mirror of
https://github.com/freqtrade/freqtrade.git
synced 2024-11-14 12:13:57 +00:00
4ca6e61726
closes #10540
586 lines
23 KiB
Python
586 lines
23 KiB
Python
"""
|
|
Dataprovider
|
|
Responsible to provide data to the bot
|
|
including ticker and orderbook data, live and historical candle (OHLCV) data
|
|
Common Interface for bot and strategy to access data.
|
|
"""
|
|
|
|
import logging
|
|
from collections import deque
|
|
from datetime import datetime, timezone
|
|
from typing import Any, Dict, List, Optional, Tuple
|
|
|
|
from pandas import DataFrame, Timedelta, Timestamp, to_timedelta
|
|
|
|
from freqtrade.configuration import TimeRange
|
|
from freqtrade.constants import (
|
|
FULL_DATAFRAME_THRESHOLD,
|
|
Config,
|
|
ListPairsWithTimeframes,
|
|
PairWithTimeframe,
|
|
)
|
|
from freqtrade.data.history import get_datahandler, load_pair_history
|
|
from freqtrade.enums import CandleType, RPCMessageType, RunMode, TradingMode
|
|
from freqtrade.exceptions import ExchangeError, OperationalException
|
|
from freqtrade.exchange import Exchange, timeframe_to_prev_date, timeframe_to_seconds
|
|
from freqtrade.exchange.types import OrderBook
|
|
from freqtrade.misc import append_candles_to_dataframe
|
|
from freqtrade.rpc import RPCManager
|
|
from freqtrade.rpc.rpc_types import RPCAnalyzedDFMsg
|
|
from freqtrade.util import PeriodicCache
|
|
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
NO_EXCHANGE_EXCEPTION = "Exchange is not available to DataProvider."
|
|
MAX_DATAFRAME_CANDLES = 1000
|
|
|
|
|
|
class DataProvider:
|
|
def __init__(
|
|
self,
|
|
config: Config,
|
|
exchange: Optional[Exchange],
|
|
pairlists=None,
|
|
rpc: Optional[RPCManager] = None,
|
|
) -> None:
|
|
self._config = config
|
|
self._exchange = exchange
|
|
self._pairlists = pairlists
|
|
self.__rpc = rpc
|
|
self.__cached_pairs: Dict[PairWithTimeframe, Tuple[DataFrame, datetime]] = {}
|
|
self.__slice_index: Optional[int] = None
|
|
self.__slice_date: Optional[datetime] = None
|
|
|
|
self.__cached_pairs_backtesting: Dict[PairWithTimeframe, DataFrame] = {}
|
|
self.__producer_pairs_df: Dict[
|
|
str, Dict[PairWithTimeframe, Tuple[DataFrame, datetime]]
|
|
] = {}
|
|
self.__producer_pairs: Dict[str, List[str]] = {}
|
|
self._msg_queue: deque = deque()
|
|
|
|
self._default_candle_type = self._config.get("candle_type_def", CandleType.SPOT)
|
|
self._default_timeframe = self._config.get("timeframe", "1h")
|
|
|
|
self.__msg_cache = PeriodicCache(
|
|
maxsize=1000, ttl=timeframe_to_seconds(self._default_timeframe)
|
|
)
|
|
|
|
self.producers = self._config.get("external_message_consumer", {}).get("producers", [])
|
|
self.external_data_enabled = len(self.producers) > 0
|
|
|
|
def _set_dataframe_max_index(self, limit_index: int):
|
|
"""
|
|
Limit analyzed dataframe to max specified index.
|
|
Only relevant in backtesting.
|
|
:param limit_index: dataframe index.
|
|
"""
|
|
self.__slice_index = limit_index
|
|
|
|
def _set_dataframe_max_date(self, limit_date: datetime):
|
|
"""
|
|
Limit infomrative dataframe to max specified index.
|
|
Only relevant in backtesting.
|
|
:param limit_date: "current date"
|
|
"""
|
|
self.__slice_date = limit_date
|
|
|
|
def _set_cached_df(
|
|
self, pair: str, timeframe: str, dataframe: DataFrame, candle_type: CandleType
|
|
) -> None:
|
|
"""
|
|
Store cached Dataframe.
|
|
Using private method as this should never be used by a user
|
|
(but the class is exposed via `self.dp` to the strategy)
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param dataframe: analyzed dataframe
|
|
:param candle_type: Any of the enum CandleType (must match trading mode!)
|
|
"""
|
|
pair_key = (pair, timeframe, candle_type)
|
|
self.__cached_pairs[pair_key] = (dataframe, datetime.now(timezone.utc))
|
|
|
|
# For multiple producers we will want to merge the pairlists instead of overwriting
|
|
def _set_producer_pairs(self, pairlist: List[str], producer_name: str = "default"):
|
|
"""
|
|
Set the pairs received to later be used.
|
|
|
|
:param pairlist: List of pairs
|
|
"""
|
|
self.__producer_pairs[producer_name] = pairlist
|
|
|
|
def get_producer_pairs(self, producer_name: str = "default") -> List[str]:
|
|
"""
|
|
Get the pairs cached from the producer
|
|
|
|
:returns: List of pairs
|
|
"""
|
|
return self.__producer_pairs.get(producer_name, []).copy()
|
|
|
|
def _emit_df(self, pair_key: PairWithTimeframe, dataframe: DataFrame, new_candle: bool) -> None:
|
|
"""
|
|
Send this dataframe as an ANALYZED_DF message to RPC
|
|
|
|
:param pair_key: PairWithTimeframe tuple
|
|
:param dataframe: Dataframe to emit
|
|
:param new_candle: This is a new candle
|
|
"""
|
|
if self.__rpc:
|
|
msg: RPCAnalyzedDFMsg = {
|
|
"type": RPCMessageType.ANALYZED_DF,
|
|
"data": {
|
|
"key": pair_key,
|
|
"df": dataframe.tail(1),
|
|
"la": datetime.now(timezone.utc),
|
|
},
|
|
}
|
|
self.__rpc.send_msg(msg)
|
|
if new_candle:
|
|
self.__rpc.send_msg(
|
|
{
|
|
"type": RPCMessageType.NEW_CANDLE,
|
|
"data": pair_key,
|
|
}
|
|
)
|
|
|
|
def _replace_external_df(
|
|
self,
|
|
pair: str,
|
|
dataframe: DataFrame,
|
|
last_analyzed: datetime,
|
|
timeframe: str,
|
|
candle_type: CandleType,
|
|
producer_name: str = "default",
|
|
) -> None:
|
|
"""
|
|
Add the pair data to this class from an external source.
|
|
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param candle_type: Any of the enum CandleType (must match trading mode!)
|
|
"""
|
|
pair_key = (pair, timeframe, candle_type)
|
|
|
|
if producer_name not in self.__producer_pairs_df:
|
|
self.__producer_pairs_df[producer_name] = {}
|
|
|
|
_last_analyzed = datetime.now(timezone.utc) if not last_analyzed else last_analyzed
|
|
|
|
self.__producer_pairs_df[producer_name][pair_key] = (dataframe, _last_analyzed)
|
|
logger.debug(f"External DataFrame for {pair_key} from {producer_name} added.")
|
|
|
|
def _add_external_df(
|
|
self,
|
|
pair: str,
|
|
dataframe: DataFrame,
|
|
last_analyzed: datetime,
|
|
timeframe: str,
|
|
candle_type: CandleType,
|
|
producer_name: str = "default",
|
|
) -> Tuple[bool, int]:
|
|
"""
|
|
Append a candle to the existing external dataframe. The incoming dataframe
|
|
must have at least 1 candle.
|
|
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param candle_type: Any of the enum CandleType (must match trading mode!)
|
|
:returns: False if the candle could not be appended, or the int number of missing candles.
|
|
"""
|
|
pair_key = (pair, timeframe, candle_type)
|
|
|
|
if dataframe.empty:
|
|
# The incoming dataframe must have at least 1 candle
|
|
return (False, 0)
|
|
|
|
if len(dataframe) >= FULL_DATAFRAME_THRESHOLD:
|
|
# This is likely a full dataframe
|
|
# Add the dataframe to the dataprovider
|
|
self._replace_external_df(
|
|
pair,
|
|
dataframe,
|
|
last_analyzed=last_analyzed,
|
|
timeframe=timeframe,
|
|
candle_type=candle_type,
|
|
producer_name=producer_name,
|
|
)
|
|
return (True, 0)
|
|
|
|
if (
|
|
producer_name not in self.__producer_pairs_df
|
|
or pair_key not in self.__producer_pairs_df[producer_name]
|
|
):
|
|
# We don't have data from this producer yet,
|
|
# or we don't have data for this pair_key
|
|
# return False and 1000 for the full df
|
|
return (False, 1000)
|
|
|
|
existing_df, _ = self.__producer_pairs_df[producer_name][pair_key]
|
|
|
|
# CHECK FOR MISSING CANDLES
|
|
# Convert the timeframe to a timedelta for pandas
|
|
timeframe_delta: Timedelta = to_timedelta(timeframe)
|
|
local_last: Timestamp = existing_df.iloc[-1]["date"] # We want the last date from our copy
|
|
# We want the first date from the incoming
|
|
incoming_first: Timestamp = dataframe.iloc[0]["date"]
|
|
|
|
# Remove existing candles that are newer than the incoming first candle
|
|
existing_df1 = existing_df[existing_df["date"] < incoming_first]
|
|
|
|
candle_difference = (incoming_first - local_last) / timeframe_delta
|
|
|
|
# If the difference divided by the timeframe is 1, then this
|
|
# is the candle we want and the incoming data isn't missing any.
|
|
# If the candle_difference is more than 1, that means
|
|
# we missed some candles between our data and the incoming
|
|
# so return False and candle_difference.
|
|
if candle_difference > 1:
|
|
return (False, int(candle_difference))
|
|
if existing_df1.empty:
|
|
appended_df = dataframe
|
|
else:
|
|
appended_df = append_candles_to_dataframe(existing_df1, dataframe)
|
|
|
|
# Everything is good, we appended
|
|
self._replace_external_df(
|
|
pair,
|
|
appended_df,
|
|
last_analyzed=last_analyzed,
|
|
timeframe=timeframe,
|
|
candle_type=candle_type,
|
|
producer_name=producer_name,
|
|
)
|
|
return (True, 0)
|
|
|
|
def get_producer_df(
|
|
self,
|
|
pair: str,
|
|
timeframe: Optional[str] = None,
|
|
candle_type: Optional[CandleType] = None,
|
|
producer_name: str = "default",
|
|
) -> Tuple[DataFrame, datetime]:
|
|
"""
|
|
Get the pair data from producers.
|
|
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param candle_type: Any of the enum CandleType (must match trading mode!)
|
|
:returns: Tuple of the DataFrame and last analyzed timestamp
|
|
"""
|
|
_timeframe = self._default_timeframe if not timeframe else timeframe
|
|
_candle_type = self._default_candle_type if not candle_type else candle_type
|
|
|
|
pair_key = (pair, _timeframe, _candle_type)
|
|
|
|
# If we have no data from this Producer yet
|
|
if producer_name not in self.__producer_pairs_df:
|
|
# We don't have this data yet, return empty DataFrame and datetime (01-01-1970)
|
|
return (DataFrame(), datetime.fromtimestamp(0, tz=timezone.utc))
|
|
|
|
# If we do have data from that Producer, but no data on this pair_key
|
|
if pair_key not in self.__producer_pairs_df[producer_name]:
|
|
# We don't have this data yet, return empty DataFrame and datetime (01-01-1970)
|
|
return (DataFrame(), datetime.fromtimestamp(0, tz=timezone.utc))
|
|
|
|
# We have it, return this data
|
|
df, la = self.__producer_pairs_df[producer_name][pair_key]
|
|
return (df.copy(), la)
|
|
|
|
def add_pairlisthandler(self, pairlists) -> None:
|
|
"""
|
|
Allow adding pairlisthandler after initialization
|
|
"""
|
|
self._pairlists = pairlists
|
|
|
|
def historic_ohlcv(self, pair: str, timeframe: str, candle_type: str = "") -> DataFrame:
|
|
"""
|
|
Get stored historical candle (OHLCV) data
|
|
:param pair: pair to get the data for
|
|
:param timeframe: timeframe to get data for
|
|
:param candle_type: '', mark, index, premiumIndex, or funding_rate
|
|
"""
|
|
_candle_type = (
|
|
CandleType.from_string(candle_type)
|
|
if candle_type != ""
|
|
else self._config["candle_type_def"]
|
|
)
|
|
saved_pair: PairWithTimeframe = (pair, str(timeframe), _candle_type)
|
|
if saved_pair not in self.__cached_pairs_backtesting:
|
|
timerange = TimeRange.parse_timerange(
|
|
None
|
|
if self._config.get("timerange") is None
|
|
else str(self._config.get("timerange"))
|
|
)
|
|
|
|
startup_candles = self.get_required_startup(str(timeframe))
|
|
tf_seconds = timeframe_to_seconds(str(timeframe))
|
|
timerange.subtract_start(tf_seconds * startup_candles)
|
|
|
|
logger.info(
|
|
f"Loading data for {pair} {timeframe} "
|
|
f"from {timerange.start_fmt} to {timerange.stop_fmt}"
|
|
)
|
|
|
|
self.__cached_pairs_backtesting[saved_pair] = load_pair_history(
|
|
pair=pair,
|
|
timeframe=timeframe,
|
|
datadir=self._config["datadir"],
|
|
timerange=timerange,
|
|
data_format=self._config["dataformat_ohlcv"],
|
|
candle_type=_candle_type,
|
|
)
|
|
return self.__cached_pairs_backtesting[saved_pair].copy()
|
|
|
|
def get_required_startup(self, timeframe: str) -> int:
|
|
freqai_config = self._config.get("freqai", {})
|
|
if not freqai_config.get("enabled", False):
|
|
return self._config.get("startup_candle_count", 0)
|
|
else:
|
|
startup_candles = self._config.get("startup_candle_count", 0)
|
|
indicator_periods = freqai_config["feature_parameters"]["indicator_periods_candles"]
|
|
# make sure the startupcandles is at least the set maximum indicator periods
|
|
self._config["startup_candle_count"] = max(startup_candles, max(indicator_periods))
|
|
tf_seconds = timeframe_to_seconds(timeframe)
|
|
train_candles = freqai_config["train_period_days"] * 86400 / tf_seconds
|
|
total_candles = int(self._config["startup_candle_count"] + train_candles)
|
|
logger.info(
|
|
f"Increasing startup_candle_count for freqai on {timeframe} to {total_candles}"
|
|
)
|
|
return total_candles
|
|
|
|
def get_pair_dataframe(
|
|
self, pair: str, timeframe: Optional[str] = None, candle_type: str = ""
|
|
) -> DataFrame:
|
|
"""
|
|
Return pair candle (OHLCV) data, either live or cached historical -- depending
|
|
on the runmode.
|
|
Only combinations in the pairlist or which have been specified as informative pairs
|
|
will be available.
|
|
:param pair: pair to get the data for
|
|
:param timeframe: timeframe to get data for
|
|
:return: Dataframe for this pair
|
|
:param candle_type: '', mark, index, premiumIndex, or funding_rate
|
|
"""
|
|
if self.runmode in (RunMode.DRY_RUN, RunMode.LIVE):
|
|
# Get live OHLCV data.
|
|
data = self.ohlcv(pair=pair, timeframe=timeframe, candle_type=candle_type)
|
|
else:
|
|
# Get historical OHLCV data (cached on disk).
|
|
timeframe = timeframe or self._config["timeframe"]
|
|
data = self.historic_ohlcv(pair=pair, timeframe=timeframe, candle_type=candle_type)
|
|
# Cut date to timeframe-specific date.
|
|
# This is necessary to prevent lookahead bias in callbacks through informative pairs.
|
|
if self.__slice_date:
|
|
cutoff_date = timeframe_to_prev_date(timeframe, self.__slice_date)
|
|
data = data.loc[data["date"] < cutoff_date]
|
|
if len(data) == 0:
|
|
logger.warning(f"No data found for ({pair}, {timeframe}, {candle_type}).")
|
|
return data
|
|
|
|
def get_analyzed_dataframe(self, pair: str, timeframe: str) -> Tuple[DataFrame, datetime]:
|
|
"""
|
|
Retrieve the analyzed dataframe. Returns the full dataframe in trade mode (live / dry),
|
|
and the last 1000 candles (up to the time evaluated at this moment) in all other modes.
|
|
:param pair: pair to get the data for
|
|
:param timeframe: timeframe to get data for
|
|
:return: Tuple of (Analyzed Dataframe, lastrefreshed) for the requested pair / timeframe
|
|
combination.
|
|
Returns empty dataframe and Epoch 0 (1970-01-01) if no dataframe was cached.
|
|
"""
|
|
pair_key = (pair, timeframe, self._config.get("candle_type_def", CandleType.SPOT))
|
|
if pair_key in self.__cached_pairs:
|
|
if self.runmode in (RunMode.DRY_RUN, RunMode.LIVE):
|
|
df, date = self.__cached_pairs[pair_key]
|
|
else:
|
|
df, date = self.__cached_pairs[pair_key]
|
|
if self.__slice_index is not None:
|
|
max_index = self.__slice_index
|
|
df = df.iloc[max(0, max_index - MAX_DATAFRAME_CANDLES) : max_index]
|
|
return df, date
|
|
else:
|
|
return (DataFrame(), datetime.fromtimestamp(0, tz=timezone.utc))
|
|
|
|
@property
|
|
def runmode(self) -> RunMode:
|
|
"""
|
|
Get runmode of the bot
|
|
can be "live", "dry-run", "backtest", "edgecli", "hyperopt" or "other".
|
|
"""
|
|
return RunMode(self._config.get("runmode", RunMode.OTHER))
|
|
|
|
def current_whitelist(self) -> List[str]:
|
|
"""
|
|
fetch latest available whitelist.
|
|
|
|
Useful when you have a large whitelist and need to call each pair as an informative pair.
|
|
As available pairs does not show whitelist until after informative pairs have been cached.
|
|
:return: list of pairs in whitelist
|
|
"""
|
|
|
|
if self._pairlists:
|
|
return self._pairlists.whitelist.copy()
|
|
else:
|
|
raise OperationalException("Dataprovider was not initialized with a pairlist provider.")
|
|
|
|
def clear_cache(self):
|
|
"""
|
|
Clear pair dataframe cache.
|
|
"""
|
|
self.__cached_pairs = {}
|
|
# Don't reset backtesting pairs -
|
|
# otherwise they're reloaded each time during hyperopt due to with analyze_per_epoch
|
|
# self.__cached_pairs_backtesting = {}
|
|
self.__slice_index = 0
|
|
|
|
# Exchange functions
|
|
|
|
def refresh(
|
|
self,
|
|
pairlist: ListPairsWithTimeframes,
|
|
helping_pairs: Optional[ListPairsWithTimeframes] = None,
|
|
) -> None:
|
|
"""
|
|
Refresh data, called with each cycle
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
final_pairs = (pairlist + helping_pairs) if helping_pairs else pairlist
|
|
# refresh latest ohlcv data
|
|
self._exchange.refresh_latest_ohlcv(final_pairs)
|
|
# refresh latest trades data
|
|
self.refresh_latest_trades(pairlist)
|
|
|
|
def refresh_latest_trades(self, pairlist: ListPairsWithTimeframes) -> None:
|
|
"""
|
|
Refresh latest trades data (if enabled in config)
|
|
"""
|
|
|
|
use_public_trades = self._config.get("exchange", {}).get("use_public_trades", False)
|
|
if use_public_trades:
|
|
if self._exchange:
|
|
self._exchange.refresh_latest_trades(pairlist)
|
|
|
|
@property
|
|
def available_pairs(self) -> ListPairsWithTimeframes:
|
|
"""
|
|
Return a list of tuples containing (pair, timeframe) for which data is currently cached.
|
|
Should be whitelist + open trades.
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
return list(self._exchange._klines.keys())
|
|
|
|
def ohlcv(
|
|
self, pair: str, timeframe: Optional[str] = None, copy: bool = True, candle_type: str = ""
|
|
) -> DataFrame:
|
|
"""
|
|
Get candle (OHLCV) data for the given pair as DataFrame
|
|
Please use the `available_pairs` method to verify which pairs are currently cached.
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param candle_type: '', mark, index, premiumIndex, or funding_rate
|
|
:param copy: copy dataframe before returning if True.
|
|
Use False only for read-only operations (where the dataframe is not modified)
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
if self.runmode in (RunMode.DRY_RUN, RunMode.LIVE):
|
|
_candle_type = (
|
|
CandleType.from_string(candle_type)
|
|
if candle_type != ""
|
|
else self._config["candle_type_def"]
|
|
)
|
|
return self._exchange.klines(
|
|
(pair, timeframe or self._config["timeframe"], _candle_type), copy=copy
|
|
)
|
|
else:
|
|
return DataFrame()
|
|
|
|
def trades(
|
|
self, pair: str, timeframe: Optional[str] = None, copy: bool = True, candle_type: str = ""
|
|
) -> DataFrame:
|
|
"""
|
|
Get candle (TRADES) data for the given pair as DataFrame
|
|
Please use the `available_pairs` method to verify which pairs are currently cached.
|
|
This is not meant to be used in callbacks because of lookahead bias.
|
|
:param pair: pair to get the data for
|
|
:param timeframe: Timeframe to get data for
|
|
:param candle_type: '', mark, index, premiumIndex, or funding_rate
|
|
:param copy: copy dataframe before returning if True.
|
|
Use False only for read-only operations (where the dataframe is not modified)
|
|
"""
|
|
if self.runmode in (RunMode.DRY_RUN, RunMode.LIVE):
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
_candle_type = (
|
|
CandleType.from_string(candle_type)
|
|
if candle_type != ""
|
|
else self._config["candle_type_def"]
|
|
)
|
|
return self._exchange.trades(
|
|
(pair, timeframe or self._config["timeframe"], _candle_type), copy=copy
|
|
)
|
|
elif self.runmode in (RunMode.BACKTEST, RunMode.HYPEROPT):
|
|
data_handler = get_datahandler(
|
|
self._config["datadir"], data_format=self._config["dataformat_trades"]
|
|
)
|
|
trades_df = data_handler.trades_load(
|
|
pair, self._config.get("trading_mode", TradingMode.SPOT)
|
|
)
|
|
return trades_df
|
|
|
|
else:
|
|
return DataFrame()
|
|
|
|
def market(self, pair: str) -> Optional[Dict[str, Any]]:
|
|
"""
|
|
Return market data for the pair
|
|
:param pair: Pair to get the data for
|
|
:return: Market data dict from ccxt or None if market info is not available for the pair
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
return self._exchange.markets.get(pair)
|
|
|
|
def ticker(self, pair: str):
|
|
"""
|
|
Return last ticker data from exchange
|
|
:param pair: Pair to get the data for
|
|
:return: Ticker dict from exchange or empty dict if ticker is not available for the pair
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
try:
|
|
return self._exchange.fetch_ticker(pair)
|
|
except ExchangeError:
|
|
return {}
|
|
|
|
def orderbook(self, pair: str, maximum: int) -> OrderBook:
|
|
"""
|
|
Fetch latest l2 orderbook data
|
|
Warning: Does a network request - so use with common sense.
|
|
:param pair: pair to get the data for
|
|
:param maximum: Maximum number of orderbook entries to query
|
|
:return: dict including bids/asks with a total of `maximum` entries.
|
|
"""
|
|
if self._exchange is None:
|
|
raise OperationalException(NO_EXCHANGE_EXCEPTION)
|
|
return self._exchange.fetch_l2_order_book(pair, maximum)
|
|
|
|
def send_msg(self, message: str, *, always_send: bool = False) -> None:
|
|
"""
|
|
Send custom RPC Notifications from your bot.
|
|
Will not send any bot in modes other than Dry-run or Live.
|
|
:param message: Message to be sent. Must be below 4096.
|
|
:param always_send: If False, will send the message only once per candle, and suppress
|
|
identical messages.
|
|
Careful as this can end up spaming your chat.
|
|
Defaults to False
|
|
"""
|
|
if self.runmode not in (RunMode.DRY_RUN, RunMode.LIVE):
|
|
return
|
|
|
|
if always_send or message not in self.__msg_cache:
|
|
self._msg_queue.append(message)
|
|
self.__msg_cache[message] = True
|