Huge News!Announcing our $40M Series B led by Abstract Ventures.Learn More
Socket
Sign inDemoInstall
Socket

saritasa-invocations

Package Overview
Dependencies
Maintainers
1
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

saritasa-invocations

Collection of invoke commands used by Saritasa

  • 1.3.0
  • PyPI
  • Socket score

Maintainers
1

saritasa-invocations

GitHub Workflow Status (with event) PyPI PyPI - Status PyPI - Python Version PyPI - License PyPI - Downloads Ruff

Collection of invoke commands used by Saritasa

Table of contents

Installation

pip install saritasa-invocations

or if you are using poetry

poetry add saritasa-invocations

Configuration

Configuration can be set in tasks.py file.

Below is an example of config:

import invoke

import saritasa_invocations

ns = invoke.Collection(
    saritasa_invocations.docker,
    saritasa_invocations.git,
    saritasa_invocations.github_actions,
    saritasa_invocations.pre_commit,
    saritasa_invocations.system,
)

# Configurations for run command
ns.configure(
    {
        "run": {
            "pty": True,
            "echo": True,
        },
        "saritasa_invocations": saritasa_invocations.Config(
            pre_commit=saritasa_invocations.PreCommitSettings(
                hooks=(
                    "pre-commit",
                    "pre-push",
                    "commit-msg",
                )
            ),
            git=saritasa_invocations.GitSettings(
                merge_ff="true",
                pull_ff="only",
            ),
            docker=saritasa_invocations.DockerSettings(
                main_containers=(
                    "opensearch",
                    "redis",
                ),
            ),
            system=saritasa_invocations.SystemSettings(
                vs_code_settings_template=".vscode/recommended_settings.json",
                settings_template="config/.env.local",
                save_settings_from_template_to="config/.env",
            ),
            # Default K8S Settings shared between envs
            k8s_defaults=saritasa_invocations.K8SDefaultSettings(
                proxy="teleport.company.com",
                db_config=saritasa_invocations.K8SDBSettings(
                    namespace="db",
                    pod_selector="app=pod-selector-db",
                ),
            )
        ),
    },
)

# For K8S settings you just need to create a instances of K8SSettings for each
# environnement. It'll be all collected automatically.
saritasa_invocations.K8SSettings(
    name="dev",
    cluster="teleport.company.somewhere.com",
    namespace="project_name",
)
saritasa_invocations.K8SSettings(
    name="prod",
    cluster="teleport.client.somewhere.com",
    namespace="project_name",
    proxy="teleport.client.com",
)

Modules

printing

While this module doesn't contain any invocations, it's used to print message via rich.panel.Panel. There are three types:

  • print_success - print message in green panel
  • print_warning - print message in yellow panel
  • print_error - print message in red panel

system

system.copy-local-settings

Copies local template for settings into specified file

Settings:

  • settings_template path to settings template (Default: config/settings/local.template.py)
  • save_settings_from_template_to path to where save settings (Default: config/settings/local.py)
system.copy-vscode-settings

Copies local template for vscode settings into .vscode folder

Settings:

  • vs_code_settings_template path to settings template (Default: .vscode/recommended_settings.json)
system.chown

Change ownership of files to user(current user by default).

Shortcut for owning apps dir by specified user after some files were generated using docker-compose (migrations, new app, etc).

system.create-tmp-folder

Create folder for temporary files(.tmp).

git

git.set-git-setting

Set git setting in config

git.setup

Preform setup of git:

  • Install pre-commit hooks
  • Set merge.ff
  • Set pull.ff

Settings:

  • merge_ff setting value for merge.ff (Default: false)
  • pull_ff setting value for pull.ff (Default: only)
git.clone-repo

Clone repo or pull latest changes to specified repo

git.blame-copy

Command for creating copies of a file with git blame history saving.

Original script written in bash here

Usage:

  inv git.blame-copy <path to original file> <path to copy>,<path to copy>...

If <path to copy> is file, then data will be copied in it.

If <path to copy> is directory, then data will be copied in provided directory with original name.

Algorithm:

  1. Remember current HEAD state
  2. For each copy path: move file to copy path, restore file using checkout, remember result commits
  3. Restore state of branch
  4. Move file to temp file
  5. Merge copy commits to branch
  6. Move file to it's original path from temp file

Settings:

  • copy_commit_template template for commits created during command workflow
  • copy_init_message_template template for init message printed at command start

Template variables:

  • action - The copy algorithm consists of several intermediate actions (creating temporary files, merging commits, etc.) The action variable stores the header of the intermediate action.
  • original_path - Contains value of first argument of the command (path of original file that will be copied)
  • destination_paths - Sequence of paths to which the original file will be copied
  • project_task - project task that will be parsed from current git branch. If no task found in branch, then will be empty

Default values for templates:

  • copy_commit_template:
  "[automated-commit]: {action}\n\n"
  "copy: {original_path}\n"
  "to:\n* {destination_paths}\n\n"
  "{project_task}"
  • copy_init_message_template:
  "Copy {original_path} to:\n"
  "* {destination_paths}\n\n"
  "Count of created commits: {commits_count}"

pre-commit

pre-commit.install

Install git hooks via pre-commit.

Settings:

  • hooks list of hooks to install (Default: ["pre-commit", "pre-push", "commit-msg"])
pre-commit.run-hooks

Run all hooks against all files.

pre-commit.update

Update pre-commit dependencies.

docker

docker.build-service

Build service image from docker compose

docker.buildpack

Build project via pack-cli

Settings:

  • buildpack_builder image tag of builder (Default: paketobuildpacks/builder:base)
  • buildpack_runner image tag of runner (Default: paketobuildpacks/run:base)
  • build_image_tag image tag of builder (Default: Name of project from project_name)
  • buildpack_requirements_path path to folder with requirements (Default: requirements)
docker.stop-all-containers

Shortcut for stopping ALL running docker containers

docker.up

Bring up main containers and start them.

Settings:

  • main_containers image tag of builder (Default: ["postgres", "redis"])
docker.stop

Stop main containers.

Settings:

  • main_containers image tag of builder (Default: ["postgres", "redis"])
docker.clear

Stop and remove all containers defined in docker-compose. Also remove images.

github-actions

github-actions.set-up-hosts

Add hosts to /etc/hosts.

Settings:

  • hosts image tag of builder (Default: see docker-main-containers)

python

As of now we support two environments for python local and docker.

  • local is a python that is located in your current virtualenv
  • docker is python that is located inside your docker image of service (python_docker_service).

This was done to have ability to run code against environment close deployed one or simply test it out.

Example of usage

PYTHON_ENV=docker inv python.run --command="--version"
python.run

Run python command depending on PYTHON_ENV variable(docker or local).

Settings:

  • entry python entry command (Default: python)
  • docker_service python service name (Default: web)
  • docker_service_params params for docker (Default: --rm)

django

django.manage

Run manage.py with specified command.

This command also handle starting of required services and waiting DB to be ready.

Requires django_probes

Settings:

  • manage_file_path path to manage.py file (Default: ./manage.py)
django.makemigrations

Run makemigrations command and chown created migrations (only for docker env).

django.check_new_migrations

Check if there is new migrations or not. Result should be check via exit code.

django.migrate

Run migrate command.

Settings:

  • migrate_command migrate command (Default: migrate)
django.resetdb

Reset database to initial state (including test DB).

Requires django-extensions

Settings:

  • settings_path default django settings (Default: config.settings.local)
django.createsuperuser

Create superuser.

Settings:

  • default_superuser_email default email of superuser. if empty, will try to grab it from git config, before resorting to default (Default: root@localhost)
  • default_superuser_username default username of superuser if empty, will try to grab it from git config, before resorting to default (Default: root)
  • default_superuser_password default password of superuser (Default: root)
  • verbose_email_name verbose name for email field (Default: Email address)
  • verbose_username_name verbose name for username field (Default: Username)
  • verbose_password_name verbose name for password field (Default: Password)

Note:

  • Values for verbose_email_name, verbose_username_name, verbose_password_name should match with verbose names of model that used this setting
django.run

Run development web-server.

Settings:

  • runserver_docker_params params for docker (Default: --rm --service-ports)
  • runserver_command runserver command (Default: runserver_plus)
  • runserver_host host of server (Default: 0.0.0.0)
  • runserver_port port of server (Default: 8000)
  • runserver_params params for runserver command (Default: "")
django.shell

Shortcut for manage.py shell command.

Settings:

  • shell_command command to start python shell (Default: shell_plus --ipython)
django.dbshell

Open database shell with credentials from current django settings.

django.recompile-messages

Generate and recompile translation messages.

Requires gettext

Settings:

  • makemessages_params params for makemessages command (Default: --all --ignore venv)
  • compilemessages_params params for compilemessages command (Default: "")
django.load-db-dump

Reset db and load db dump.

Uses resetdb and load-db-dump

Settings:

  • django_settings_path default django settings (Default: config.settings.local)
django.backup-local-db

Back up local db.

Uses backup_local_db

Settings:

  • settings_path default django settings (Default: config.settings.local)
django.backup-remote-db

Make dump of remote db and download it.

Uses create_dump and get-dump

Settings:

  • settings_path default django settings (Default: config.settings.local)

  • remote_db_config_mapping Mapping of db config Default:

    {
        "dbname": "RDS_DB_NAME",
        "host": "RDS_DB_HOST",
        "port": "RDS_DB_PORT",
        "username": "RDS_DB_USER",
        "password": "RDS_DB_PASSWORD",
    }
    
django.load-remote-db

Make dump of remote db and download it and apply to local db.

Uses create_dump and get-dump and load-db-dump

Settings:

  • settings_path default django settings (Default: config.settings.local)
django.startapp

Create django app from a template using cookiecutter.

Settings:

  • app_boilerplate_link link to app template
  • app_template_directory path to app template in project template (Default: .)
  • apps_path path to apps folder in project (Default: apps)
django.wait-for-database

Launch docker compose and wait for database connection.

fastapi

fastapi.run

Run development web-server.

Settings:

  • docker_params params for docker (Default: --rm --service-ports)
  • uvicorn_command uvicorn command (Default: -m uvicorn)
  • app path to fastapi app (Default: config:fastapi_app)
  • host host of server (Default: 0.0.0.0)
  • port port of server (Default: 8000)
  • params params for uvicorn (Default: --reload)

alembic

alembic.run

Run alembic command

Settings:

  • command alembic command (Default: -m alembic)
  • connect_attempts numbers of attempts to connect to database (Default: 10)
alembic.autogenerate

Generate migrations

Settings:

  • migrations_folder migrations files location (Default: db/migrations/versions)
alembic.upgrade

Upgrade database

alembic.downgrade

Downgrade database

alembic.check-for-migrations

Check if there any missing migrations to be generated

alembic.check-for-adjust-messages

Check migration files for adjust messages

Settings:

  • migrations_folder migrations files location (Default: db/migrations/versions)
  • adjust_messages list of alembic adjust messages (Default: # ### commands auto generated by Alembic - please adjust! ###, # ### end Alembic commands ###)
alembic.load-db-dump

Reset db and load db dump.

Uses downgrade and load-db-dump

Requires python-decouple

Installed with [env_settings]

Settings:

  • db_config_mapping Mapping of db config

    Default:

    {
      "dbname": "rds_db_name",
      "host": "rds_db_host",
      "port": "rds_db_port",
      "username": "rds_db_user",
      "password": "rds_db_password",
    }
    
alembic.backup-local-db

Back up local db.

Uses backup_local_db

Requires python-decouple

Installed with [env_settings]

Settings:

  • db_config_mapping Mapping of db config

    Default:

    {
      "dbname": "rds_db_name",
      "host": "rds_db_host",
      "port": "rds_db_port",
      "username": "rds_db_user",
      "password": "rds_db_password",
    }
    
alembic.backup-remote-db

Make dump of remote db and download it.

Uses create_dump and get-dump

Requires python-decouple

Installed with [env_settings]

Settings:

  • db_config_mapping Mapping of db config

    Default:

    {
      "dbname": "rds_db_name",
      "host": "rds_db_host",
      "port": "rds_db_port",
      "username": "rds_db_user",
      "password": "rds_db_password",
    }
    
alembic.load-remote-db

Make dump of remote db and download it and apply to local db.

Uses create-dump and get-dump and load-db-dump

Requires python-decouple

Installed with [env_settings]

Settings:

  • db_config_mapping Mapping of db config

    Default:

    {
      "dbname": "rds_db_name",
      "host": "rds_db_host",
      "port": "rds_db_port",
      "username": "rds_db_user",
      "password": "rds_db_password",
    }
    
alembic.wait-for-database

Launch docker compose and wait for database connection.

celery

celery.run

Start celery worker.

Settings:

  • app path to app (Default: config.celery.app)
  • scheduler scheduler (Default: django)
  • loglevel log level for celery (Default: info)
  • extra_params extra params for worker (Default: ("--beat",))
  • local_cmd command for celery (Default: celery --app {app} worker --scheduler={scheduler} --loglevel={info} {extra_params})
  • service_name name of celery service (Default: celery)
celery.send-task

Send task to celery worker.

Settings:

  • app path to app (Default: config.celery.app)

open-api

open-api.validate-swagger

Check that generated open_api spec is valid. This command uses drf-spectacular and it's default validator. It creates spec file in ./tmp folder and then validates it.

db

db.load-db-dump

Load db dump to local db.

Settings:

  • load_dump_command template for load command(Default located in _config.pp > dbSettings)
  • dump_filename filename for dump (Default: local_db_dump)
  • load_additional_params additional params for load command (Default: --quite)
db.backup-local-db

Back up local db.

Settings:

  • dump_command template for dump command (Default located in _config.pp > dbSettings)
  • dump_filename filename for dump (Default: local_db_dump)
  • dump_additional_params additional params for dump command (Default: --no-owner)

k8s

For K8S settings you just need to create a instances of K8SSettings for each environnement. It'll be all collected automatically.

k8s.login

Login into k8s via teleport.

Settings:

  • proxy teleport proxy (REQUIRED)
  • port teleport port (Default: 443)
  • auth teleport auth method (Default: github)
k8s.set-context

Set k8s context to current project

Settings:

  • namespace namespace for k8s (Default: Name of project from project_name)
k8s.logs

Get logs for k8s pod

Settings:

  • default_component default component (Default: backend)
k8s.pods

Get pods from k8s.

k8s.execute

Execute command inside k8s pod.

Settings:

  • default_component default component (Default: backend)
  • default_entry default entry cmd (Default: /cnb/lifecycle/launcher bash)
k8s.python-shell

Enter python shell inside k8s pod.

Settings:

  • default_component default component (Default: backend)
  • python_shell shell cmd (Default: shell_plus)
k8s.health-check

Check health of component.

Settings:

  • default_component default component (Default: backend)
  • health_check health check cmd (Default: health_check)
k8s.download-file

Download file from pod.

  • default_component default component (Default: backend)

db-k8s

While you probably won't use this module directly some other modules commands are use it(getting remote db dump)

Make sure to set up these configs:

  • pod_namespace db namespace (REQUIRED)
  • pod_selector pod selector for db (REQUIRED)
db-k8s.create-dump

Execute dump command in db pod.

Settings:

  • pod_namespace db namespace (REQUIRED)
  • pod_selector pod selector for db (REQUIRED)
  • get_pod_name_command template for fetching db pod (Default located in _config.pp > K8SdbSettings)
  • dump_filename default dump filename (Default: Name of project from project_name plus _db_dump)
  • dump_command dump command template (Default located in _config.pp > K8SDBSettings)
  • dump_dir folder where to put dump file (Default: tmp)
  • dump_additional_params additional dump commands (Default: --no-owner)
db-k8s.get-dump

Download db data from db pod if it present

Settings:

  • pod_namespace db namespace (REQUIRED)
  • pod_selector pod selector for db (REQUIRED)
  • get_pod_name_command template for fetching db pod (Default located in _config.pp > K8SDBSettings)
  • dump_filename default dump filename (Default: Name of project from project_name plus _db_dump)

cruft

Cruft is a tool used to synchronize changes with cookiecutter based boilerplates.

cruft.check-for-cruft-files

Check that there are no cruft files (*.rej).

cruft.create_project

Not invocation, but a shortcut for creating cruft projects for testing boilerplates

poetry

poetry.install

Install dependencies via poetry.

poetry.update

Update dependencies with respect to version constraints using poetry up plugin.

Fallbacks to poetry update in case of an error.

poetry.update-to-latest

Update dependencies to latest versions using poetry up plugin.

By default fallbacks to update task in case of an error. Use --no-fallback to stop on error.

pip

pip.install

Install dependencies via pip.

Settings:

  • dependencies_folder path to folder with dependencies files (Default: requirements)
pip.compile

Compile dependencies via pip-compile.

Settings:

  • dependencies_folder path to folder with dependencies files (Default: requirements)
  • in_files sequence of .in files (Default: "production.in", "development.in")

mypy

mypy.run

Run mypy in path with params.

Settings:

  • mypy_entry python entry command (Default: -m mypy)

pytest

pytest.run

Run pytest in path with params.

Settings:

  • pytest_entry python entry command (Default: -m pytest)

secrets

secrets.setup-env-credentials

Fill specified credentials in your file from k8s. This invocations downloads .env file from pod in k8s. It will replace specified credentials(--credentials) in specified file .env file (--env_file_path or .env as default)

Requires python-decouple

Settings for k8s:

  • secret_file_path_in_pod path to secret in pod (REQUIRED)
  • temp_secret_file_path path for temporary file (Default: .env.to_delete)

Keywords

FAQs


Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

  • Package Alerts
  • Integrations
  • Docs
  • Pricing
  • FAQ
  • Roadmap
  • Changelog

Packages

npm

Stay in touch

Get open source security insights delivered straight into your inbox.


  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc