Socket
Book a DemoInstallSign in
Socket

artifician

Package Overview
Dependencies
Maintainers
1
Alerts
File Explorer

Advanced tools

Socket logo

Install Socket

Detect and block malicious and high-risk dependencies

Install

artifician

Artifician is an event driven framework developed to simplify the process of preparation of the dataset for Artificial Intelligence models.

0.6.4
Source
pipPyPI
Maintainers
1

codecov CI/CD GitHub release (latest by date) PyPI version GitBook License Conda


Logo

Artifician

Artifician is an event driven library developed to simplify and speed up the process of preparation of the datasets for Artificial Intelligence models.


Getting Started

Pre-requisites

  • python v3.6 or later

Installation

Binary installers for the latest released version are available at the Python Package Index (PyPI) and on Conda

# or PyPI
pip install artifician
# conda
conda install -c plato_solutions artifician

Documentation

Please visit Aritfician Docs

Usage

from artifician.dataset import *
from artifician.feature_definition import *
from artifician.processors.normalizer import *

  
def extract_domain_name(sample):  
    """function for extracting the path from the given URL"""
    domain_name = sample.split("//")[-1].split('/')[0] 
 
    return domain_name  
 
input_data = ['https://www.google.com/', 'https://www.youtube.com/']  
  
dataset = Dataset() # initializing dataset object
url_domain = FeatureDefinition(extract_domain_name, dataset) # initializing feature_definition and passing extractor function name as a parameter and subscribing it to dataset
normalizer = Normalizer(PropertiesNormalizer(), url_domain delimiter = {'delimiter': ["."]})  # Initializing normalizer (processor) and passing properties normalizer as a parameter and subscribing it to url_domain
  
  
""" Now we are all set to go, all we have to do is call add_samples method on the dataset object and pass the input data
after calling the add_samples, url_domain will start its execution and extract the data using extract_domain_name function, as soon url_domain
feature is processed normalizer will start it execution and furthur is will process the data extracted by url_domain. The processed data is then
passed back to the dataset. Following diagram will make it more clear for you. """ 

prepared_data = dataset.add_samples(input_data)  
print(prepared_data)  
  

Output

01
0https://www.google.com/[(www, 0), (google, 1), (com, 2)]
1https://www.youtube.com/[(www, 0), (youtube, 1), (com, 2)]

Keywords

python artifician

FAQs

Did you know?

Socket

Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.

Install

Related posts

SocketSocket SOC 2 Logo

Product

About

Packages

Stay in touch

Get open source security insights delivered straight into your inbox.

  • Terms
  • Privacy
  • Security

Made with ⚡️ by Socket Inc

U.S. Patent No. 12,346,443 & 12,314,394. Other pending.