Security News
Fluent Assertions Faces Backlash After Abandoning Open Source Licensing
Fluent Assertions is facing backlash after dropping the Apache license for a commercial model, leaving users blindsided and questioning contributor rights.
Ollama Chat is a conversational AI chat client that uses Ollama to interact with local large language models (LLMs).
To get up and running with Ollama Chat follows these steps:
Install and start Ollama
Install Ollama Chat
pip install ollama-chat
To update Ollama Chat:
pip install -U ollama-chat
To start Ollama Chat, open a terminal prompt and run the Ollama Chat application:
ollama-chat
A web browser is launched and opens the Ollama Chat web application.
By default, a configuration file, "ollama-chat.json", is created in the user's home directory.
To start a conversation from the command line, use the -m
argument:
ollama-chat -m "Why is the sky blue?"
To start a named template from the command line, use the -t
and -v
arguments:
ollama-chat -t AskAristotle -v Subject "Why is the sky blue?"
Conversation Templates allow you to repeat the same prompts with different models. Templates can define variables for
use in the template title and prompt text (e.g., {{var}}
).
There are two ways to create a template. Click "Add Template" from the index page, and a new template is created and opened in the template editor. The other way is to click "Template" from a conversation view's menu.
Ollama Chat supports special prompt commands that allow you to include files, images, and URL content in your prompt, among other things. The following prompt commands are available:
/file
- include a file
/file README.md
Please summarize the README file.
/image
- include an image
/image image.jpeg
Please summarize the image.
/dir
- include files from a directory
/dir src/ollama_chat py
Please provide a summary for each Ollama Chat source file.
/url
- include a URL resource
/url https://craigahobbs.github.io/ollama-chat/README.md
Please summarize the README file.
/do
- execute a conversation template by name
/do city-report -v CityState "Seattle, WA"
To get prompt command help use the -h
option:
/file -h
This package is developed using python-build. It was started using python-template as follows:
template-specialize python-template/template/ ollama-chat/ -k package ollama-chat -k name 'Craig A. Hobbs' -k email 'craigahobbs@gmail.com' -k github 'craigahobbs' -k noapi 1
FAQs
An Ollama chat web application
We found that ollama-chat demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
Fluent Assertions is facing backlash after dropping the Apache license for a commercial model, leaving users blindsided and questioning contributor rights.
Research
Security News
Socket researchers uncover the risks of a malicious Python package targeting Discord developers.
Security News
The UK is proposing a bold ban on ransomware payments by public entities to disrupt cybercrime, protect critical services, and lead global cybersecurity efforts.