🚀 Big News: Socket Acquires Coana to Bring Reachability Analysis to Every Appsec Team.Learn more

clip-bbox

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

clip-bbox

Python library for detecting image objects with natural language text labels

0.3.1

Unpopular package

Quality

This package is not very popular.

Found 1 instance in 1 package

Uses eval

Supply chain risk

Package uses dynamic code execution (e.g., eval()), which is a dangerous practice. This can prevent the code from running in certain environments and increases the risk that the code may contain exploits or malicious behavior.

Found 1 instance in 1 package

Network access

Supply chain risk

This module accesses the network.

Found 1 instance in 1 package

Shell access

Supply chain risk

This module accesses the system shell. Accessing the system shell increases the risk of executing arbitrary code.

Found 1 instance in 1 package

Dependencies have 7 high and critical alerts.

Critical CVE and High CVE

Maintainers
1

CLIP_BBox

CLIP_BBox is a Python library for detecting image objects with natural language text labels.

Build Status codecov GitHub GitHub issues PyPI Documentation Status

Overview / About

CLIP is a neural network, pretrained on image-text pairs, that can predict the most relevant text snippet for a given image.

Given an image and a natural language text label, CLIP_BBox will obtain the image's spatial embedding and text label's embedding from CLIP, compute the similarity heatmap between the embeddings, then draw bounding boxes around the image regions with the highest image-text correspondences.

Note

The files for building the CLIP model (clip.py, model.py, newpad.py, simple_tokenizer.py) are third-party code from the CLIP repo. They are not included in test coverage.

Features

The library provides functions for the following operations:

  • Getting and appropriately reshaping an image's spatial embedding from the CLIP model before it performs attention-pooling
  • Getting a text snippet's embedding from CLIP
  • Computing the similarity heatmap between an image's spatial and text embeddings from CLIP
  • Drawing bounding boxes on an image, given a similarity heatmap

Install

Use pip to install clip_bbox as a Python package:

$ pip install clip-bbox

Usage Examples

Command Line Script

usage: python -m clip_bbox [-h] imgpath caption outpath

positional arguments:
  imgpath     path to input image
  caption     caption of input image
  outpath     path to output image displaying bounding boxes

optional arguments:
  -h, --help  show this help message and exit

To draw bounding boxes on an image based on its caption, run

$ python -m clip_bbox "path/to/img.png" "caption of your image" "path/to/output_path.png"

Python Module

To draw bounding boxes on an image based on its caption, do the following:

from clip_bbox import run_clip_bbox

run_clip_bbox("path/to/img.png", "caption of your image", "path/to/output_path.png")

Example Output

Here is an example output image for the caption "a camera on a tripod":

example output

FAQs

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts