Socket
Socket
Sign inDemoInstall

pai-repl

Package Overview
Dependencies
3
Maintainers
1
Alerts
File Explorer

Install Socket

Detect and block malicious and high-risk dependencies

Install

    pai-repl

Python REPL with LLM integration


Maintainers
1

Readme

pai: A Python REPL with a built in AI agent.

It's like OpenAI's Code Interpreter running in your Python REPL.

A fully functional Python REPL with a built in AI agent that can generate and run code using the history as context.

PyPI - Version PyPI - Python Version

Features

  • Full Python REPL
  • Built-in AI agent that can generate and run code in the REPL
  • REPL history is used for LLM context.
  • Edit and approve all generated code before it is executed
  • Runs locally on your machine, so it has full system and internet access
  • Supports OpenAI and llama.cpp

Demo

Installation

pip install pai-repl

Usage

OpenAI models

The default model is OpenAI GPT-4. You will need to set your OpenAI API key.

$ export OPENAI_API_KEY=<your key>
$ pai

Specify OpenAI model

$ pai --openai gpt-3.5-turbo

llama.cpp compatible models

llama-cpp-python is an optional dependency because it requires native libraries to be installed so it must be installed using the llama extra.

$ pip install pai-repl[llama]
$ pai --llama <path to model>

Starting the REPL

When you invoke pai, it will start an interactive Python REPL.

$ pai
INP> print('howdy')
OUT> howdy

Start the agent

Start the agent with pai: <prompt>.

This will generate code using the prompt and REPL history. You can accept, edit or cancel the code. Immediately after the generated code is run, the LLM is called again with the new REPL history. This loop continues until the task is complete or you cancel the agent with Ctrl+C.

INP> pai: list files in the current directory
LLM>
import os
os.listdir()

OK?> import os
    ...> os.listdir()

One off code generation

Generate code with gen: <prompt>.

The generated code will be displayed and you can accept, edit or cancel it. Unlike the pai command, the LLM is not called again after the code is run.

INP> nums = [1,2,3]
INP> gen: average nums
LLM>
# to find the average of the numbers, we sum all the elements and then divide by the number of elements

average_nums = sum(nums) / len(nums)
average_nums

OK?> # to find the average of the numbers, we sum all the elements and then
    ...> divide by the number of elements
    ...>
    ...> average_nums = sum(nums) / len(nums)
    ...> average_nums
OUT> 2.0
INP>

REPL features

reset() will reset the REPL state and history. This is useful if you want to start a new task or want to start over. No previous history will be used for LLM context.

INP> a = 1
INP> reset()
INP> a
OUT> Traceback (most recent call last):
  File "<console>", line 1, in <module>
NameError: name 'a' is not defined
INP>

Run shell commands with !

INP> !ls
OUT> README.md
     assets
     setup.py
     ...

Quickstart from the command line

You can prompt pai from the command line

$ pai "find the largest file in the current directory"

FAQs


Did you know?

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc