The Lean CLI is a cross-platform CLI aimed at making it easier to develop with the LEAN engine locally and in the cloud.
Visit the documentation website for comprehensive and up-to-date documentation.
- Project scaffolding
- Local autocomplete
- Local data downloading
- Local backtesting
- Local debugging
- Local research environment
- Local optimization
- Local live trading
- Local backtest report creation
- Cloud synchronization
- Cloud backtesting
- Cloud optimization
- Cloud live trading
The CLI can be installed and updated by running pip install --upgrade lean
.
Note that many commands in the CLI require Docker to run. See Get Docker for instructions on how to install Docker for your operating system.
After installing the CLI, open a terminal in an empty directory and run lean init
. This command downloads the latest configuration file and sample data from the QuantConnect/Lean repository. We recommend running all Lean CLI commands in the same directory lean init
was ran in.
The Lean CLI supports multiple workflows. The examples below serve as a starting point, you're free to mix local and cloud features in any way you'd like.
A cloud-focused workflow (local development, cloud execution) with the CLI may look like this:
- Open a terminal in the directory you ran
lean init
in. - Run
lean cloud pull
to pull remotely changed files. - Start programming locally and run backtests in the cloud with
lean cloud backtest "Project Name" --open --push
whenever there is something to backtest. The--open
flag means that the backtest results will be opened in the browser when done, while the--push
flag means that local changes are pushed to the cloud before running the backtest. - Whenever you want to create a new project, run
lean create-project "Project Name"
andlean cloud push --project "Project Name"
to create a new project containing some basic code and to push it to the cloud. - When you're finished for the moment, run
lean cloud push
to push all locally changed files to the cloud.
A locally-focused workflow (local development, local execution) with the CLI may look like this:
- Open a terminal in the directory you ran
lean init
in. - Run
lean create-project "Project Name"
to create a new project with some basic code to get you started. - Work on your strategy in
./Project Name
. - Run
lean research "Project Name"
to start a Jupyter Lab session to perform research in. - Run
lean backtest "Project Name"
to run a backtest whenever there's something to test. This runs your strategy in a Docker container containing the same packages as the ones used on QuantConnect.com, but with your own data.
Note: the readme only contains the --help
text of all commands. Visit the documentation website for more comprehensive documentation.
lean backtest
lean build
lean cloud backtest
lean cloud live
lean cloud optimize
lean cloud pull
lean cloud push
lean cloud status
lean config get
lean config list
lean config set
lean config unset
lean create-project
lean data download
lean data generate
lean gui logs
lean gui restart
lean gui start
lean gui stop
lean init
lean library add
lean library remove
lean live
lean login
lean logout
lean logs
lean optimize
lean report
lean research
lean whoami
Backtest a project locally using Docker.
Usage: lean backtest [OPTIONS] PROJECT
Backtest a project locally using Docker.
If PROJECT is a directory, the algorithm in the main.py or Main.cs file inside it will be executed.
If PROJECT is a file, the algorithm in the specified file will be executed.
Go to the following url to learn how to debug backtests locally using the Lean CLI:
https://www.lean.io/docs/lean-cli/backtesting/debugging
By default the official LEAN engine image is used. You can override this using the --image option. Alternatively you
can set the default engine image for all commands using `lean config set engine-image <image>`.
Options:
--output DIRECTORY Directory to store results in (defaults to PROJECT/backtests/TIMESTAMP)
-d, --detach Run the backtest in a detached Docker container and return immediately
--debug [pycharm|ptvsd|vsdbg|rider]
Enable a certain debugging method (see --help for more information)
--data-provider [Local|QuantConnect|Terminal Link]
Update the Lean configuration file to retrieve data from the given provider
--download-data Update the Lean configuration file to download data from the QuantConnect API, alias
for --data-provider QuantConnect
--data-purchase-limit INTEGER The maximum amount of QCC to spend on downloading data during the backtest when using
QuantConnect as data provider
--release Compile C# projects in release configuration instead of debug
--image TEXT The LEAN engine image to use (defaults to quantconnect/lean:latest)
--update Pull the LEAN engine image before running the backtest
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/backtest.py
Build Docker images of your own version of LEAN and the Alpha Streams SDK.
Usage: lean build [OPTIONS] [ROOT]
Build Docker images of your own version of LEAN and the Alpha Streams SDK.
ROOT must point to a directory containing the LEAN repository and the Alpha Streams SDK repository:
https://github.com/QuantConnect/Lean & https://github.com/QuantConnect/AlphaStreams
When ROOT is not given, the current directory is used as root directory.
This command performs the following actions:
1. The lean-cli/foundation:latest image is built from Lean/DockerfileLeanFoundation(ARM).
2. LEAN is compiled in a Docker container using the lean-cli/foundation:latest image.
3. The Alpha Streams SDK is compiled in a Docker container using the lean-cli/foundation:latest image.
4. The lean-cli/engine:latest image is built from Lean/Dockerfile using lean-cli/foundation:latest as base image.
5. The lean-cli/research:latest image is built from Lean/DockerfileJupyter using lean-cli/engine:latest as base image.
6. The default engine image is set to lean-cli/engine:latest.
7. The default research image is set to lean-cli/research:latest.
When the foundation Dockerfile is the same as the official foundation Dockerfile, quantconnect/lean:foundation is
used instead of building a custom foundation image.
Options:
--tag TEXT The tag to apply to custom images (defaults to latest)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/build.py
Backtest a project in the cloud.
Usage: lean cloud backtest [OPTIONS] PROJECT
Backtest a project in the cloud.
PROJECT must be the name or id of the project to run a backtest for.
If the project that has to be backtested has been pulled to the local drive with `lean cloud pull` it is possible to
use the --push option to push local modifications to the cloud before running the backtest.
Options:
--name TEXT The name of the backtest (a random one is generated if not specified)
--push Push local modifications to the cloud before running the backtest
--open Automatically open the results in the browser when the backtest is finished
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/backtest.py
Start live trading for a project in the cloud.
Usage: lean cloud live [OPTIONS] PROJECT
Start live trading for a project in the cloud.
PROJECT must be the name or the id of the project to start live trading for.
By default an interactive wizard is shown letting you configure the deployment. If --brokerage is given the command
runs in non-interactive mode. In this mode the CLI does not prompt for input or confirmation. In non-interactive
mode the options specific to the given brokerage are required, as well as --node, --auto-restart, --notify-order-
events and --notify-insights.
Options:
--brokerage [Paper Trading|Interactive Brokers|Tradier|OANDA|Bitfinex|Coinbase Pro|Binance|Kraken|FTX|Zerodha|Samco]
The brokerage to use
--ib-user-name TEXT Your Interactive Brokers username
--ib-account TEXT Your Interactive Brokers account id
--ib-password TEXT Your Interactive Brokers password
--ib-data-feed BOOLEAN Whether the Interactive Brokers price data feed must be used instead of the
QuantConnect price data feed
--tradier-account-id TEXT Your Tradier account id
--tradier-access-token TEXT Your Tradier access token
--tradier-environment [demo|real]
The environment to run in, demo for the Developer Sandbox, real for live trading
--oanda-account-id TEXT Your OANDA account id
--oanda-access-token TEXT Your OANDA API token
--oanda-environment [demo|real]
The environment to run in, demo for fxTrade Practice, real for fxTrade
--bitfinex-api-key TEXT Your Bitfinex API key
--bitfinex-api-secret TEXT Your Bitfinex API secret
--gdax-api-key TEXT Your Coinbase Pro API key
--gdax-api-secret TEXT Your Coinbase Pro API secret
--gdax-passphrase TEXT Your Coinbase Pro API passphrase
--gdax-environment [paper|live]
The environment to run in, paper for the sandbox, live for live trading
--binance-api-key TEXT Your Binance API key
--binance-api-secret TEXT Your Binance API secret
--binance-environment [demo|real]
The environment to run in, demo for testnet, real for the production environment
--kraken-api-key TEXT Your Kraken API key
--kraken-api-secret TEXT Your Kraken API secret
--kraken-verification-tier TEXT
Your Kraken Verification Tier
--ftx-api-key TEXT Your FTX API key
--ftx-api-secret TEXT Your FTX API secret
--ftx-account-tier TEXT Your FTX Account Tier
--ftx-exchange-name TEXT FTX exchange name [FTX, FTXUS]
--zerodha-api-key TEXT Your Kite Connect API key
--zerodha-access-token TEXT Your Kite Connect access token
--zerodha-product-type [MIS|CNC|NRML]
MIS if you are targeting intraday products, CNC if you are targeting delivery
products, NRML if you are targeting carry forward products
--zerodha-trading-segment [EQUITY|COMMODITY]
EQUITY if you are trading equities on NSE or BSE, COMMODITY if you are trading
commodities on MCX
--samco-client-id TEXT Your Samco account Client ID
--samco-client-password TEXT Your Samco account password
--samco-year-of-birth TEXT Your year of birth (YYYY) registered with Samco
--samco-product-type [MIS|CNC|NRML]
MIS if you are targeting intraday products, CNC if you are targeting delivery
products, NRML if you are targeting carry forward products
--samco-trading-segment [EQUITY|COMMODITY]
EQUITY if you are trading equities on NSE or BSE, COMMODITY if you are trading
commodities on MCX
--node TEXT The name or id of the live node to run on
--auto-restart BOOLEAN Whether automatic algorithm restarting must be enabled
--notify-order-events BOOLEAN Whether notifications must be sent for order events
--notify-insights BOOLEAN Whether notifications must be sent for emitted insights
--notify-emails TEXT A comma-separated list of 'email:subject' pairs configuring email-notifications
--notify-webhooks TEXT A comma-separated list of 'url:HEADER_1=VALUE_1:HEADER_2=VALUE_2:etc' pairs
configuring webhook-notifications
--notify-sms TEXT A comma-separated list of phone numbers configuring SMS-notifications
--push Push local modifications to the cloud before starting live trading
--open Automatically open the live results in the browser once the deployment starts
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/live.py
Optimize a project in the cloud.
Usage: lean cloud optimize [OPTIONS] PROJECT
Optimize a project in the cloud.
PROJECT must be the name or id of the project to optimize.
An interactive prompt will be shown to configure the optimizer. If --target is given the command runs in non-
interactive mode. In this mode the CLI does not prompt for input and the following options become required:
--target, --target-direction, --parameter, --node and --parallel-nodes.
In non-interactive mode the --parameter option can be provided multiple times to configure multiple parameters:
- --parameter <name> <min value> <max value> <step size>
- --parameter my-first-parameter 1 10 0.5 --parameter my-second-parameter 20 30 5
In non-interactive mode the --constraint option can be provided multiple times to configure multiple constraints:
- --constraint "<statistic> <operator> <value>"
- --constraint "Sharpe Ratio >= 0.5" --constraint "Drawdown < 0.25"
If the project that has to be optimized has been pulled to the local drive with `lean cloud pull` it is possible to
use the --push option to push local modifications to the cloud before running the optimization.
Options:
--target TEXT The target statistic of the optimization
--target-direction [min|max] Whether the target must be minimized or maximized
--parameter <TEXT FLOAT FLOAT FLOAT>...
The 'parameter min max step' pairs configuring the parameters to optimize
--constraint TEXT The 'statistic operator value' pairs configuring the constraints of the optimization
--node [O2-8|O4-12|O8-16] The node type to run the optimization on
--parallel-nodes INTEGER The number of nodes that may be run in parallel
--name TEXT The name of the optimization (a random one is generated if not specified)
--push Push local modifications to the cloud before starting the optimization
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/optimize.py
Pull projects from QuantConnect to the local drive.
Usage: lean cloud pull [OPTIONS]
Pull projects from QuantConnect to the local drive.
This command overrides the content of local files with the content of their respective counterparts in the cloud.
This command will not delete local files for which there is no counterpart in the cloud.
Options:
--project TEXT Name or id of the project to pull (all cloud projects if not specified)
--pull-bootcamp Pull Boot Camp projects (disabled by default)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/pull.py
Push local projects to QuantConnect.
Usage: lean cloud push [OPTIONS]
Push local projects to QuantConnect.
This command overrides the content of cloud files with the content of their respective local counterparts.
This command will not delete cloud files which don't have a local counterpart.
Options:
--project DIRECTORY Path to the local project to push (all local projects if not specified)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/push.py
Show the live trading status of a project in the cloud.
Usage: lean cloud status [OPTIONS] PROJECT
Show the live trading status of a project in the cloud.
PROJECT must be the name or the id of the project to show the status for.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/cloud/status.py
Get the current value of a configurable option.
Usage: lean config get [OPTIONS] KEY
Get the current value of a configurable option.
Sensitive options like credentials cannot be retrieved this way for security reasons. Please open
~/.lean/credentials if you want to see your currently stored credentials.
Run `lean config list` to show all available options.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/config/get.py
List the configurable options and their current values.
Usage: lean config list [OPTIONS]
List the configurable options and their current values.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/config/list.py
Set a configurable option.
Usage: lean config set [OPTIONS] KEY VALUE
Set a configurable option.
Run `lean config list` to show all available options.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/config/set.py
Unset a configurable option.
Usage: lean config unset [OPTIONS] KEY
Unset a configurable option.
Run `lean config list` to show all available options.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/config/unset.py
Create a new project containing starter code.
Usage: lean create-project [OPTIONS] NAME
Create a new project containing starter code.
If NAME is a path containing subdirectories those will be created automatically.
The default language can be set using `lean config set default-language python/csharp`.
Options:
-l, --language [python|csharp] The language of the project to create
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/create_project.py
Purchase and download data from QuantConnect Datasets.
Usage: lean data download [OPTIONS]
Purchase and download data from QuantConnect Datasets.
An interactive wizard will show to walk you through the process of selecting data, accepting the CLI API Access and
Data Agreement and payment. After this wizard the selected data will be downloaded automatically.
If --dataset is given the command runs in non-interactive mode. In this mode the CLI does not prompt for input or
confirmation but only halts when the agreement must be accepted. In non-interactive mode all options specific to the
selected dataset as well as --organization are required.
See the following url for the data that can be purchased and downloaded with this command:
https://www.quantconnect.com/datasets
Options:
--dataset TEXT The name of the dataset to download non-interactively
--organization TEXT The name or id of the organization to purchase and download data with
--overwrite Overwrite existing local data
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/data/download.py
Generate random market data.
Usage: lean data generate [OPTIONS]
Generate random market data.
This uses the random data generator in LEAN to generate realistic market data using a Brownian motion model.
This generator supports the following security types, tick types and resolutions:
| Security type | Generated tick types | Supported resolutions |
| ------------- | -------------------- | ------------------------------------ |
| Equity | Trade | Tick, Second, Minute, Hour and Daily |
| Forex | Quote | Tick, Second, Minute, Hour and Daily |
| CFD | Quote | Tick, Second, Minute, Hour and Daily |
| Future | Trade and Quote | Tick, Second, Minute, Hour and Daily |
| Crypto | Trade and Quote | Tick, Second, Minute, Hour and Daily |
| Option | Trade and Quote | Minute |
The following data densities are available:
- Dense: at least one data point per resolution step.
- Sparse: at least one data point per 5 resolution steps.
- VerySparse: at least one data point per 50 resolution steps.
Example which generates minute data for 100 equity symbols since 2015-01-01:
$ lean data generate --start=20150101 --symbol-count=100
Example which generates daily data for 100 crypto symbols since 2015-01-01:
$ lean data generate --start=20150101 --symbol-count=100 --security-type=Crypto --resolution=Daily
By default the official LEAN engine image is used. You can override this using the --image option. Alternatively you
can set the default engine image for all commands using `lean config set engine-image <image>`.
Options:
--start [yyyyMMdd] Start date for the data to generate in yyyyMMdd format [required]
--end [yyyyMMdd] End date for the data to generate in yyyyMMdd format (defaults to today)
--symbol-count INTEGER RANGE The number of symbols to generate data for [required]
--security-type [Equity|Forex|Cfd|Future|Crypto|Option]
The security type to generate data for (defaults to Equity)
--resolution [Tick|Second|Minute|Hour|Daily]
The resolution of the generated data (defaults to Minute)
--data-density [Dense|Sparse|VerySparse]
The density of the generated data (defaults to Dense)
--include-coarse BOOLEAN Whether coarse universe data should be generated for Equity data (defaults to True)
--market TEXT The market to generate data for (defaults to standard market for the security type)
--image TEXT The LEAN engine image to use (defaults to quantconnect/lean:latest)
--update Pull the LEAN engine image before running the generator
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/data/generate.py
See the logs of the local GUI.
Usage: lean gui logs [OPTIONS]
See the logs of the local GUI.
Options:
-f, --follow Update the logs in real-time while the GUI is running
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/gui/logs.py
Restart the local GUI and open it in the browser.
Usage: lean gui restart [OPTIONS]
Restart the local GUI and open it in the browser.
Options:
--no-open Skip opening the local GUI in the browser after restarting it
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/gui/restart.py
Start the local GUI.
Usage: lean gui start [OPTIONS]
Start the local GUI.
Options:
--organization TEXT The name or id of the organization with the local GUI module subscription
--port INTEGER The port to run the local GUI on (defaults to 5612)
--no-open Skip opening the local GUI in the browser after starting it
--shortcut Create a desktop shortcut for launching the local GUI
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/gui/start.py
Stop the local GUI.
Usage: lean gui stop [OPTIONS]
Stop the local GUI.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/gui/stop.py
Scaffold a Lean configuration file and data directory.
Usage: lean init [OPTIONS]
Scaffold a Lean configuration file and data directory.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/init.py
Add a custom library to a project.
Usage: lean library add [OPTIONS] PROJECT NAME
Add a custom library to a project.
PROJECT must be the path to the project.
NAME must be the name of a NuGet package (for C# projects) or of a PyPI package (for Python projects).
If --version is not given, the package is pinned to the latest compatible version. For C# projects, this is the
latest available version. For Python projects, this is the latest version compatible with Python 3.6 (which is what
the Docker images use).
Custom C# libraries are added to your project's .csproj file, which is then restored if dotnet is on your PATH and
the --no-local flag has not been given.
Custom Python libraries are added to your project's requirements.txt file and are installed in your local Python
environment so you get local autocomplete for the library. The last step can be skipped with the --no-local flag.
C# example usage:
$ lean library add "My CSharp Project" Microsoft.ML
$ lean library add "My CSharp Project" Microsoft.ML --version 1.5.5
Python example usage:
$ lean library add "My Python Project" tensorflow
$ lean library add "My Python Project" tensorflow --version 2.5.0
Options:
--version TEXT The version of the library to add (defaults to latest compatible version)
--no-local Skip making changes to your local environment
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/library/add.py
Remove a custom library from a project.
Usage: lean library remove [OPTIONS] PROJECT NAME
Remove a custom library from a project.
PROJECT must be the path to the project directory.
NAME must be the name of the NuGet package (for C# projects) or of the PyPI package (for Python projects) to remove.
Custom C# libraries are removed from the project's .csproj file, which is then restored if dotnet is on your PATH
and the --no-local flag has not been given.
Custom Python libraries are removed from the project's requirements.txt file.
C# example usage:
$ lean library remove "My CSharp Project" Microsoft.ML
Python example usage:
$ lean library remove "My Python Project" tensorflow
Options:
--no-local Skip making changes to your local environment
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/library/remove.py
Start live trading a project locally using Docker.
Usage: lean live [OPTIONS] PROJECT
Start live trading a project locally using Docker.
If PROJECT is a directory, the algorithm in the main.py or Main.cs file inside it will be executed.
If PROJECT is a file, the algorithm in the specified file will be executed.
By default an interactive wizard is shown letting you configure the brokerage and data feed to use. If
--environment, --brokerage or --data-feed are given the command runs in non-interactive mode. In this mode the CLI
does not prompt for input.
If --environment is given it must be the name of a live environment in the Lean configuration.
If --brokerage and --data-feed are given, the options specific to the given brokerage/data feed must also be given.
The Lean config is used as fallback when a brokerage/data feed-specific option hasn't been passed in. If a required
option is not given and cannot be found in the Lean config the command aborts.
By default the official LEAN engine image is used. You can override this using the --image option. Alternatively you
can set the default engine image for all commands using `lean config set engine-image <image>`.
Options:
--environment TEXT The environment to use
--output DIRECTORY Directory to store results in (defaults to PROJECT/live/TIMESTAMP)
-d, --detach Run the live deployment in a detached Docker container and return immediately
--gui Enable monitoring and controlling of the deployment via the local GUI
--gui-organization TEXT The name or id of the organization with the local GUI module subscription
--brokerage [Paper Trading|Interactive Brokers|Tradier|OANDA|Bitfinex|Coinbase Pro|Binance|Zerodha|Samco|Terminal Link|Atreyu|Trading Technologies|Kraken|FTX]
The brokerage to use
--data-feed [Interactive Brokers|Tradier|OANDA|Bitfinex|Coinbase Pro|Binance|Zerodha|Samco|Terminal Link|Trading Technologies|Custom data only|Kraken|FTX|IQFeed]
The data feed to use
--ib-organization TEXT The name or id of the organization with the Interactive Brokers module subscription
--ib-user-name TEXT Your Interactive Brokers username
--ib-account TEXT Your Interactive Brokers account id
--ib-password TEXT Your Interactive Brokers password
--ib-enable-delayed-streaming-data BOOLEAN
Whether delayed data may be used when your algorithm subscribes to a security you
don't have a market data subscription for
--tradier-account-id TEXT Your Tradier account id
--tradier-access-token TEXT Your Tradier access token
--tradier-use-sandbox BOOLEAN Whether the developer sandbox should be used
--oanda-account-id TEXT Your OANDA account id
--oanda-access-token TEXT Your OANDA API token
--oanda-environment [Practice|Trade]
The environment to run in, Practice for fxTrade Practice, Trade for fxTrade
--bitfinex-api-key TEXT Your Bitfinex API key
--bitfinex-api-secret TEXT Your Bitfinex API secret
--gdax-api-key TEXT Your Coinbase Pro API key
--gdax-api-secret TEXT Your Coinbase Pro API secret
--gdax-passphrase TEXT Your Coinbase Pro API passphrase
--gdax-use-sandbox BOOLEAN Whether the sandbox should be used
--binance-organization TEXT The name or id of the organization with the Binance module subscription
--binance-api-key TEXT Your Binance API key
--binance-api-secret TEXT Your Binance API secret
--binance-use-testnet BOOLEAN Whether the testnet should be used
--zerodha-organization TEXT The name or id of the organization with the zerodha module subscription
--zerodha-api-key TEXT Your Kite Connect API key
--zerodha-access-token TEXT Your Kite Connect access token
--zerodha-product-type [MIS|CNC|NRML]
MIS if you are targeting intraday products, CNC if you are targeting delivery
products, NRML if you are targeting carry forward products
--zerodha-trading-segment [EQUITY|COMMODITY]
EQUITY if you are trading equities on NSE or BSE, COMMODITY if you are trading
commodities on MCX
--zerodha-history-subscription BOOLEAN
Whether you have a history API subscription for Zerodha
--samco-organization TEXT The name or id of the organization with the samco module subscription
--samco-client-id TEXT Your Samco account Client ID
--samco-client-password TEXT Your Samco account password
--samco-year-of-birth TEXT Your year of birth (YYYY) registered with Samco
--samco-product-type [MIS|CNC|NRML]
MIS if you are targeting intraday products, CNC if you are targeting delivery
products, NRML if you are targeting carry forward products
--samco-trading-segment [EQUITY|COMMODITY]
EQUITY if you are trading equities on NSE or BSE, COMMODITY if you are trading
commodities on MCX
--iqfeed-iqconnect FILE The path to the IQConnect binary
--iqfeed-username TEXT Your IQFeed username
--iqfeed-password TEXT Your IQFeed password
--iqfeed-product-name TEXT The product name of your IQFeed developer account
--iqfeed-version TEXT The product version of your IQFeed developer account
--terminal-link-organization TEXT
The name or id of the organization with the Terminal Link module subscription
--bloomberg-environment [Production|Beta]
The environment to run in
--bloomberg-server-host TEXT The host of the Bloomberg server
--bloomberg-server-port INTEGER
The port of the Bloomberg server
--bloomberg-symbol-map-file FILE
The path to the Bloomberg symbol map file
--bloomberg-emsx-broker TEXT The EMSX broker to use
--bloomberg-emsx-user-time-zone TEXT
The EMSX user timezone to use
--bloomberg-emsx-account TEXT The EMSX account to use
--bloomberg-emsx-strategy TEXT The EMSX strategy to use
--bloomberg-emsx-notes TEXT The EMSX notes to use
--bloomberg-emsx-handling TEXT The EMSX handling to use
--bloomberg-allow-modification BOOLEAN
Whether modification is allowed
--atreyu-organization TEXT The name or id of the organization with the Atreyu module subscription
--atreyu-host TEXT The host of the Atreyu server
--atreyu-req-port INTEGER The Atreyu request port
--atreyu-sub-port INTEGER The Atreyu subscribe port
--atreyu-username TEXT Your Atreyu username
--atreyu-password TEXT Your Atreyu password
--atreyu-client-id TEXT Your Atreyu client id
--atreyu-broker-mpid TEXT The broker MPID to use
--atreyu-locate-rqd TEXT The locate rqd to use
--tt-organization TEXT The name or id of the organization with the Trading Technologies module subscription
--tt-user-name TEXT Your Trading Technologies username
--tt-session-password TEXT Your Trading Technologies session password
--tt-account-name TEXT Your Trading Technologies account name
--tt-rest-app-key TEXT Your Trading Technologies REST app key
--tt-rest-app-secret TEXT Your Trading Technologies REST app secret
--tt-rest-environment TEXT The REST environment to run in
--tt-market-data-sender-comp-id TEXT
The market data sender comp id to use
--tt-market-data-target-comp-id TEXT
The market data target comp id to use
--tt-market-data-host TEXT The host of the market data server
--tt-market-data-port TEXT The port of the market data server
--tt-order-routing-sender-comp-id TEXT
The order routing sender comp id to use
--tt-order-routing-target-comp-id TEXT
The order routing target comp id to use
--tt-order-routing-host TEXT The host of the order routing server
--tt-order-routing-port TEXT The port of the order routing server
--tt-log-fix-messages BOOLEAN Whether FIX messages should be logged
--kraken-organization TEXT The name or id of the organization with the kraken module subscription
--kraken-api-key TEXT Your Kraken API key
--kraken-api-secret TEXT Your Kraken API secret
--kraken-verification-tier TEXT
Your Kraken Verification Tier
--ftx-organization TEXT The name or id of the organization with the FTX module subscription
--ftx-api-key TEXT Your FTX API key
--ftx-api-secret TEXT Your FTX API secret
--ftx-account-tier TEXT Your FTX Account Tier
--ftx-exchange-name TEXT FTX exchange name [FTX, FTXUS]
--release Compile C# projects in release configuration instead of debug
--image TEXT The LEAN engine image to use (defaults to quantconnect/lean:latest)
--update Pull the LEAN engine image before starting live trading
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/live.py
Log in with a QuantConnect account.
Usage: lean login [OPTIONS]
Log in with a QuantConnect account.
If user id or API token is not provided an interactive prompt will show.
Credentials are stored in ~/.lean/credentials and are removed upon running `lean logout`.
Options:
-u, --user-id TEXT QuantConnect user id
-t, --api-token TEXT QuantConnect API token
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/login.py
Log out and remove stored credentials.
Usage: lean logout [OPTIONS]
Log out and remove stored credentials.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/logout.py
Display the most recent backtest/live/optimization logs.
Usage: lean logs [OPTIONS]
Display the most recent backtest/live/optimization logs.
Options:
--backtest Display the most recent backtest logs (default)
--live Display the most recent live logs
--optimization Display the most recent optimization logs
--project DIRECTORY The project to get the most recent logs from
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/logs.py
Optimize a project's parameters locally using Docker.
Usage: lean optimize [OPTIONS] PROJECT
Optimize a project's parameters locally using Docker.
If PROJECT is a directory, the algorithm in the main.py or Main.cs file inside it will be executed.
If PROJECT is a file, the algorithm in the specified file will be executed.
By default an interactive wizard is shown letting you configure the optimizer. If --optimizer-config or --strategy
is given the command runs in non-interactive mode. In this mode the CLI does not prompt for input.
The --optimizer-config option can be used to specify the configuration to run the optimizer with.
When using the option it should point to a file like this (the algorithm-* properties should be omitted):
https://github.com/QuantConnect/Lean/blob/master/Optimizer.Launcher/config.json
If --strategy is given the optimizer configuration is read from the given options. In this case --strategy,
--target, --target-direction and --parameter become required.
In non-interactive mode the --parameter option can be provided multiple times to configure multiple parameters:
- --parameter <name> <min value> <max value> <step size>
- --parameter my-first-parameter 1 10 0.5 --parameter my-second-parameter 20 30 5
In non-interactive mode the --constraint option can be provided multiple times to configure multiple constraints:
- --constraint "<statistic> <operator> <value>"
- --constraint "Sharpe Ratio >= 0.5" --constraint "Drawdown < 0.25"
By default the official LEAN engine image is used. You can override this using the --image option. Alternatively you
can set the default engine image for all commands using `lean config set engine-image <image>`.
Options:
--output DIRECTORY Directory to store results in (defaults to PROJECT/optimizations/TIMESTAMP)
-d, --detach Run the optimization in a detached Docker container and return immediately
--optimizer-config FILE The optimizer configuration file that should be used
--strategy [Grid Search|Euler Search]
The optimization strategy to use
--target TEXT The target statistic of the optimization
--target-direction [min|max] Whether the target must be minimized or maximized
--parameter <TEXT FLOAT FLOAT FLOAT>...
The 'parameter min max step' pairs configuring the parameters to optimize
--constraint TEXT The 'statistic operator value' pairs configuring the constraints of the optimization
--release Compile C# projects in release configuration instead of debug
--image TEXT The LEAN engine image to use (defaults to quantconnect/lean:latest)
--update Pull the LEAN engine image before running the optimizer
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/optimize.py
Generate a report of a backtest.
Usage: lean report [OPTIONS]
Generate a report of a backtest.
This runs the LEAN Report Creator in Docker to generate a polished, professional-grade report of a backtest.
If --backtest-results is not given, a report is generated for the most recent local backtest.
The name, description, and version are optional and will be blank if not given.
If the given backtest data source file is stored in a project directory (or one of its subdirectories, like the
default <project>/backtests/<timestamp>), the default name is the name of the project directory and the default
description is the description stored in the project's config.json file.
By default the official LEAN engine image is used. You can override this using the --image option. Alternatively you
can set the default engine image for all commands using `lean config set engine-image <image>`.
Options:
--backtest-results FILE Path to the JSON file containing the backtest results
--live-results FILE Path to the JSON file containing the live trading results
--report-destination FILE Path where the generated report is stored as HTML (defaults to ./report.html)
-d, --detach Run the report creator in a detached Docker container and return immediately
--strategy-name TEXT Name of the strategy, will appear at the top-right corner of each page
--strategy-version TEXT Version number of the strategy, will appear next to the project name
--strategy-description TEXT Description of the strategy, will appear under the 'Strategy Description' section
--overwrite Overwrite --report-destination if it already contains a file
--image TEXT The LEAN engine image to use (defaults to quantconnect/lean:latest)
--update Pull the LEAN engine image before running the report creator
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/report.py
Run a Jupyter Lab environment locally using Docker.
Usage: lean research [OPTIONS] PROJECT
Run a Jupyter Lab environment locally using Docker.
By default the official LEAN research image is used. You can override this using the --image option. Alternatively
you can set the default research image using `lean config set research-image <image>`.
Options:
--port INTEGER The port to run Jupyter Lab on (defaults to 8888)
--data-provider [Local|QuantConnect|Terminal Link]
Update the Lean configuration file to retrieve data from the given provider
--download-data Update the Lean configuration file to download data from the QuantConnect API, alias
for --data-provider QuantConnect
--data-purchase-limit INTEGER The maximum amount of QCC to spend on downloading data during the research session
when using QuantConnect as data provider
-d, --detach Run Jupyter Lab in a detached Docker container and return immediately
--no-open Don't open the Jupyter Lab environment in the browser after starting it
--image TEXT The LEAN research image to use (defaults to quantconnect/research:latest)
--update Pull the LEAN research image before starting the research environment
--lean-config FILE The Lean configuration file that should be used (defaults to the nearest lean.json)
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/research.py
Display who is logged in.
Usage: lean whoami [OPTIONS]
Display who is logged in.
Options:
--verbose Enable debug logging
--help Show this message and exit.
See code: lean/commands/whoami.py
To work on the Lean CLI, clone the repository, enter an environment containing Python 3.6+ and run pip install -r requirements.txt
. This command will install the required dependencies and installs the CLI in editable mode. This means you'll be able to edit the code and immediately see the results the next time you run lean
.
If you need to add dependencies, first update setup.py
(if it is a production dependency) or requirements.txt
(if it is a development dependency) and then re-run pip install -r requirements.txt
.
The automated tests can be ran by running pytest
. The filesystem and HTTP requests are mocked when running tests to make sure they run in an isolated environment.
Can build the lean CLI by running python setup.py sdist bdist_wheel
from the root of the project and to install it pip install --force-reinstall dist/lean-dev-py3-none-any.whl
. To update the commands reference part of the readme run python scripts/readme.py
from the root of the project, after installing the new version.
Maintainers can publish new releases by pushing a Git tag containing the new version to GitHub. This will trigger a GitHub Actions workflow which releases the current master
branch to PyPI with the value of the tag as version. Make sure the version is not prefixed with "v".