Research
Security News
Malicious npm Package Targets Solana Developers and Hijacks Funds
A malicious npm package targets Solana developers, rerouting funds in 2% of transactions to a hardcoded address.
Target Dependent Sentiment Analysis (TDSA) framework. The paper associated with this repository is the following:
@InProceedings{C18-1097,
author = "Moore, Andrew
and Rayson, Paul",
title = "Bringing replication and reproduction together with generalisability in NLP: Three reproduction studies for Target Dependent Sentiment Analysis",
booktitle = "Proceedings of the 27th International Conference on Computational Linguistics",
year = "2018",
publisher = "Association for Computational Linguistics",
pages = "1132--1144",
location = "Santa Fe, New Mexico, USA",
url = "http://aclweb.org/anthology/C18-1097"
}
pip install bella-tdsa
docker run -p 9000:9000 -d --rm mooreap/corenlp
docker run -p 8000:8000 -d --rm mooreap/tweeboparserdocker
If you want to use the moses tokeniser that has been taken from the Moses project the following will need to be installed:
python -m nltk.downloader perluniprops
python -m nltk.downloader nonbreaking_prefixes
The docker Stanford and Tweebo server are only required if you are going to use the TDParse methods/models or if you are going to use any of the Stanford Tools else you do not need them.
To stop the docker servers running:
NOTE Both of these servers will run with as many threads as your machine has CPUs to limit this do the following:
docker run -p 9000:9000 -d --rm mooreap/corenlp -threads 6
will run it with 6 threadsdocker run -p 8000:8000 -d --rm mooreap/tweeboparserdocker --threads 6
will run it with 6 threadsAll of the dataset are required to be downloaded and are not stored in this repository. We recomend using the config file to state where the datasets are stored like we did but this is not a requirement as you can state where they are stored explictly in the code. For more details on the datasets and downloading them see the dataset notebook The datasets used:
NOTE Before using Mitchell and YouTuBean datasets please go through these pre-processing notebooks: Mitchell YouTuBean for splitting their data and also in Mitchell case which train test split to use.
These lexicons are required to be downloaded if you use any methods that require them. Please see the use of the config file for storing the location of the lexicons:
All the word vectors are automatically downloaded for you and they are stored in the root directory called .Bella/Vectors
which is created in your user directory e.g. on Linux that would be ~/.Bella/Vectors/
. The word vectors included in this repository are the following:
The model zoo can be found on the Git Lab repository here.
These models can be automatically downloaded through the code like the word vectors and stored in the .Bella/Models
directory which is automatically placed in your home directory for instance on Linux that would be ~/.Bella/Models
. An example of how to download and use a model is shown below:
from bella import helper
from bella.models.target import TargetDep
target_dep = helper.download_model(TargetDep, 'SemEval 14 Restaurant')
test_example_multi = [{'text' : 'This bread is tasty but the sauce is too rich', 'target': 'sauce',
'spans': [(28, 33)]}]
target_dep.predict(test_example_multi)
This example will download the Target Dependent model which is from Vo and Zhang paper that has been trained on the SemEval 2014 Resturant data and predict the sentiment of sauce from that example. As you can see the example is not simple as it has two different sentiments within the same sentence with two targets; 1. bread with a positive sentiment and 2. sauce which has a negative sentiment of which that target is the one being predicted for in this example.
To see a more in depth guide to the pre-trained models and output from them go to this notebook.
Can be found here
The best order to look at the notebooks is first look at the data with this notebook. Then looking at the notebook that describes how to load and use the saved models from the model zoo. Then go and explore the rest if you would like:
Both Tweebo and Stanford by default will run on your localhost and port 8000 and 9000 respectively by default. If you would like to run them on a different port or hostname you can change the .Bella/config.yaml
file which is created in your local home directory the first time you run something successfully through the Stanford or Tweebo tools. The file once created which is only done automatically for you when you successfully run something through the Stanford or Tweebo tools will look like this which is a yaml formatted file:
tweebo_parser:
hostname: 0.0.0.0
port: 7000
stanford_core_nlp:
hostname: http://localhost
port: 8000
If you want Tweebo or Stanford to run on a different port or hostname just change this file. For instance the example shown above is different to default as Stanford is running on port 8000 and not 9000 and Tweebo is running on port 7000 instead of 8000.
If you would like the tools to run on different hostname and port from the start without having to successfully run them through the tools before hand just create this file .Bella/config.yaml
in your local home directory with the same structure as the example but with the hostname and port you want to use.
As we use a lot of tools we list here if any of the tools that we use are licensed under a different license to that of this repository:
FAQs
Target Dependent Sentiment Analysis (TDSA) framework.
We found that bella-tdsa demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Research
Security News
A malicious npm package targets Solana developers, rerouting funds in 2% of transactions to a hardcoded address.
Security News
Research
Socket researchers have discovered malicious npm packages targeting crypto developers, stealing credentials and wallet data using spyware delivered through typosquats of popular cryptographic libraries.
Security News
Socket's package search now displays weekly downloads for npm packages, helping developers quickly assess popularity and make more informed decisions.