fire-chat
Overview
This project provides a command-line interface (CLI) for interacting with various large language models (LLMs) using the
LiteLLM wrapper. It supports multiple providers, including OpenAI, Anthropic, Azure, and Gemini. The CLI allows users to
chat with these models, manage budgets, and handle API keys efficiently.
Configuration
The configuration is managed through a $HOME/.config/fire-chat/config.yaml
file. The first time you run the CLI run.
You can copy paste the starting config file config.yaml to the location, adds your API key,
and quick start the application fire-chat
.
Installation and Usage
-
Install the CLI:
pip install --user fire-chat # requires python 3.10+
-
Configure the CLI:
Edit the $HOME/.config/fire-chat/config.yaml
file to set your preferred provider, model, and other settings.
-
Run the CLI:
fire-chat
or run with arguments (overriding config yaml file)
fire-chat --model=gpt-4o
for full list of configs, see main.py.
-
Exit:
To exit the CLI, Ctrl+C
.