Research
Security News
Malicious npm Package Targets Solana Developers and Hijacks Funds
A malicious npm package targets Solana developers, rerouting funds in 2% of transactions to a hardcoded address.
Big Data Quality, a set of tools/function that help you every day assert quality of datasets you have processed or ingested
BDQ stands for Big Data Quality, a set of tools/function that help you every day assert quality of datasets you have processed or ingested. Library leverages power of spark, hence it processes your quality checks at scale offered by spark and databricks.
Library over time evolved into DAG executor of arbitrary python/pyspark functions. This allows to scale execution with dependency tracking to next level.
On databricks run %pip install bdq==x.y.z
. Make sure you tag version number you are confortable with to ensure API stability.
This package is currently in EXPERIMENTAL stage and newer releses might change APIs (function names, parameters, etc..).
Development and testing has been performed on 12.2 LTS databricks runtime. Databricks runtime 10.4 LTS should also work with exception of SparkPipeline.step_spark_for_each_batch
and SparkPipeline.spark_metric
which require 12.2 LTS runtime to work.
bdq utilizes python logging
module, hence for the best logging experience configure logger, example setup:
import logging
import sys
# py4j is very chatty, no need to deal with it unless it's critical
logging.getLogger('py4j').setLevel(logging.CRITICAL)
# run everything else on DEBUG by default
# DEBUG prints a lot of usefull info, INFO is good in general use cases
logging.basicConfig(
stream=sys.stderr,
level=logging.DEBUG,
format='%(asctime)s %(levelname)s %(threadName)s [%(name)s] %(message)s'
)
See examples bellow for short listing of major functionalities. See tests
folder for detailed examples. Tests are meant to double as real use cases hence they serve as documentation of examples for now.
import bdq
from datetime import datetime
import pyspark.sql.functions as F
df2 = spark.createDataFrame([
[ 1, 1, "Grzegorz", datetime(2018, 1, 1), datetime(2018, 1, 1, 12, 34, 56), 26.9, 123234234345, True],
[ 3, 1, "Mike", datetime(2019, 1, 1), datetime(2018, 3, 1, 12, 34, 56), 46.7, 5667888989, False],
[ 2, 2, "Timmy", datetime(2018, 1, 1), datetime(2018, 2, 1, 12, 34, 56), 36.7, 8754857845, True]
], schema)
df2_changed = df2 \
.drop('first_login_dt') \
.withColumn('new_data', F.lit(None).cast('date')) \
.withColumn('likes', F.col('likes').cast('int'))
assert bdq.compare_schemas(df2.schema, df2_changed.schema) == {
'added': {'first_login_dt'},
'removed': {'new_data'},
'changed': {
'likes': {'before': 'bigint', 'after': 'int'}
},
'not_changed': {
'name', 'id1', 'last_login_ts', 'id2', 'credits', 'active'
}
}
import bdq
df1 = spark.createDataFrame([
[ 1, 1, "Grzegorz", datetime(2017, 1, 1), datetime(2018, 1, 1, 12, 34, 56), 26.7, 123234234345, True],
[ 2, 1, "Tim", datetime(2018, 1, 1), datetime(2018, 2, 1, 12, 34, 56), 36.7, 54545, True],
[ 3, 1, "Mike", datetime(2019, 1, 1), datetime(2018, 3, 1, 12, 34, 56), 46.7, 5667888989, False]
], schema)
df2 = spark.createDataFrame([
[ 1, 1, "Grzegorz", datetime(2018, 1, 1), datetime(2018, 1, 1, 12, 34, 56), 26.9, 123234234345, True],
[ 3, 1, "Mike", datetime(2019, 1, 1), datetime(2018, 3, 1, 12, 34, 56), 46.7, 5667888989, False],
[ 2, 2, "Timmy", datetime(2018, 1, 1), datetime(2018, 2, 1, 12, 34, 56), 36.7, 8754857845, True]
], schema)
#returns dicts with added, removed, changed and not changed dataframes, and record counts
df_diff = bdq.compare_dataframes(df1, df2, ['id1', 'id2'], True)
#show all info about compare results
bdq.display_compare_dataframes_results(df_diff)
>> Added records count: 1
>> +---+---+-----+--------------+-------------------+-------+----------+------+
>> |id1|id2|name |first_login_dt|last_login_ts |credits|likes |active|
>> +---+---+-----+--------------+-------------------+-------+----------+------+
>> |2 |2 |Timmy|2018-01-01 |2018-02-01 12:34:56|36.7 |8754857845|true |
>> +---+---+-----+--------------+-------------------+-------+----------+------+
>>
>> Removed records count: 1
>> +---+---+----+--------------+-------------------+-------+-----+------+
>> |id1|id2|name|first_login_dt|last_login_ts |credits|likes|active|
>> +---+---+----+--------------+-------------------+-------+-----+------+
>> |2 |1 |Tim |2018-01-01 |2018-02-01 12:34:56|36.7 |54545|true |
>> +---+---+----+--------------+-------------------+-------+-----+------+
>>
>> Changed records count: 1
>> +---+---+---------------------------------------------------------------------+
>> |id1|id2|changed |
>> +---+---+---------------------------------------------------------------------+
>> |1 |1 |{first_login_dt -> {2017-01-01, 2018-01-01}, credits -> {26.7, 26.9}}|
>> +---+---+---------------------------------------------------------------------+
Not changed records count: 1
with support of optional uppercasing and trimming of key columns
from datetime import datetime
from bdq.functions import surrogate_key_hash, surrogate_key_string
schema = "id1:long,id2:long,name:string,likes:int"
df1 = spark.createDataFrame([
[ 1, 1, "GrzeGorz", 1 ],
[ 1, 1, "Grzegorz", 1 ],
[ 1, 1, "Grzegorz ", 1 ],
[ 1, None, "Grzegorz", 1],
[ 1, None, None, 1],
[ 2, 1, "Tim", 1 ],
[ 3, 1, "Mike", 1 ]
], schema)
sk_df = df1.select(
'*',
surrogate_key_hash(['id1', 'id2', 'name'], rtrim=True).alias('hash'),
surrogate_key_string(['id1', 'id2', 'name'], rtrim=True).alias('hash_input')
)
>> +---+----+--------------+-----+-------------------------------------------------------------+---------------------------+
>> |id1|id2 |name |likes|hash |hash_input |
>> +---+----+--------------+-----+-------------------------------------------------------------+---------------------------+
>> |1 |1 |GrzeGorz |1 |[6F 21 99 99 4C F2 93 56 2E 7C C3 29 F9 6A 42 2F 6D 62 EC 4B]|[1, 1, GRZEGORZ] |
>> |1 |1 |Grzegorz |1 |[6F 21 99 99 4C F2 93 56 2E 7C C3 29 F9 6A 42 2F 6D 62 EC 4B]|[1, 1, GRZEGORZ] |
>> |1 |1 |Grzegorz |1 |[6F 21 99 99 4C F2 93 56 2E 7C C3 29 F9 6A 42 2F 6D 62 EC 4B]|[1, 1, GRZEGORZ] |
>> |1 |null|Grzegorz |1 |[4E 9E 3A 2B C8 37 90 A0 6A 26 9A FE 7A 78 8B CA 99 34 68 1C]|[1, @~<null>~@, GRZEGORZ] |
>> |1 |null|null |1 |[44 FF 88 2E 2A 28 3B EE EE 53 C8 7D 0F 92 D1 68 3E A2 F2 E5]|[1, @~<null>~@, @~<null>~@]|
>> |2 |1 |Tim |1 |[A8 7A B0 29 94 9F 5C D1 78 D4 28 89 8B F1 75 CD AC 70 B6 61]|[2, 1, TIM] |
>> |3 |1 |Mike |1 |[9D 6C 76 C5 17 D0 44 46 47 1A 71 ED 4A 08 75 84 1D 5C 45 09]|[3, 1, MIKE] |
>> +---+----+--------------+-----+-------------------------------------------------------------+---------------------------+
with support of showing sample data from fact table that did not satisfy integrity
from bdq import fact_dim_broken_relationship
from bdq.functions import surrogate_key_hash, surrogate_key_string
fact_df = spark.createDataFrame([
[ "Grzegorz", "IT", "EU" ],
[ "Mark", "IT", "EU" ],
[ "Justin", "IT ", "EU " ],
[ "Alice1", "IT", "USA" ],
[ "Alice2", "IT", "USA" ],
[ "Alice3", "IT", "USA" ],
[ "Alice4", "IT", "USA" ],
[ "Alice5", "IT", "USA" ],
[ "Bob", "Sales", "USA" ],
[ "Bob2", "Sales", "USA" ],
[ "Bob3", "Sales", "USA" ]
], "Name:string,Dept:string,Country:string") \
.withColumn("dept_fk", surrogate_key_hash(["Dept", "Country"], rtrim=True))
dim_df = spark.createDataFrame([
["IT", "EU", "European IT department"],
["Sales", "USA", "USA Sales dept."]
], "department:string,cntry:string,comment:string") \
.withColumn("dept_pk", surrogate_key_hash(["department", "cntry"], rtrim=True))
#direct column comparison, hence upper, lower, extra spaces affect the results
broken_df = fact_dim_broken_relationship(
fact_df=fact_df,
fk_columns=["Dept", "Country"],
dim_df=dim_df,
pk_columns=["department", "cntry"],
sample_broken_records=2
)
>> +----+-------+------------------------------------------------------------------------------------+
>> |Dept|Country|sample_records |
>> +----+-------+------------------------------------------------------------------------------------+
>> |IT |USA |[{IT, USA, Alice1, ��xz4q_��\ra�-��Vz}, {IT, USA, Alice2, ��xz4q_��\ra�-��Vz}]|
>> |IT |EU |[{IT , EU , Justin, ��H'V4�\rY,g�I�����}] |
>> +----+-------+------------------------------------------------------------------------------------+
# using surrogate keys will yield better results, because extra spaces are trimmed
broken_sk_df = fact_dim_broken_relationship(
fact_df, ["dept_fk"],
dim_df, ["dept_pk"],
sample_broken_records=2
)
>> +----+-------+------------------------------------------------------------------------------------+
>> |Dept|Country|sample_records |
>> +----+-------+------------------------------------------------------------------------------------+
>> |IT |USA |[{IT, USA, Alice1, ��xz4q_��\ra�-��Vz}, {IT, USA, Alice2, ��xz4q_��\ra�-��Vz}]|
>> +----+-------+------------------------------------------------------------------------------------+
with optional primary key conflict resolution, where there are multiple records being candidate for latest record, but they have different attribuets and there is no way of determining which one is the latest.
from datetime import datetime as dt
from bdq import get_latest_records, get_latest_records_with_pk_confict_detection_flag
increment_data = [
(1, dt(2000, 1, 1, 0, 0, 0), "1001")
,(1, dt(2000, 1, 1, 2, 0, 0), "1002")
,(2, dt(2000, 1, 1, 0, 0, 0), "2001") #carbon copy duplicate
,(2, dt(2000, 1, 1, 0, 0, 0), "2001") #carbon copy duplicate
,(3, dt(2000, 1, 1, 0, 0, 0), "3001")
,(3, dt(2000, 1, 1, 2, 0, 0), "3002#1") #pk conflict, two entries with the same technical fields, but different atributes, which one to chose?
,(3, dt(2000, 1, 1, 2, 0, 0), "3002#2") #pk conflict
]
df = spark.createDataFrame(increment_data, "pk:int,change_ts:timestamp,attr:string")
df.show(truncate=True)
>> +---+-------------------+------+
>> | pk| change_ts| attr|
>> +---+-------------------+------+
>> | 1|2000-01-01 00:00:00| 1001|
>> | 1|2000-01-01 02:00:00| 1002|
>> | 2|2000-01-01 00:00:00| 2001|
>> | 2|2000-01-01 00:00:00| 2001|
>> | 3|2000-01-01 00:00:00| 3001|
>> | 3|2000-01-01 02:00:00|3002#1|
>> | 3|2000-01-01 02:00:00|3002#2|
>> +---+-------------------+------+
primary_key_columns = ['pk']
order_by_columns = ['change_ts']
# will randomly chose one pk in case of conflict, without any notification about conflicts
simple_get_latest_df = get_latest_records(df, primary_key_columns, order_by_columns)
simple_get_latest_df.show(truncate=True)
>> +---+-------------------+------+
>> | pk| change_ts| attr|
>> +---+-------------------+------+
>> | 1|2000-01-01 02:00:00| 1002|
>> | 2|2000-01-01 00:00:00| 2001|
>> | 3|2000-01-01 02:00:00|3002#1|
>> +---+-------------------+------+
# will flag records which cause conflicts and cannot be resolved
# bad records need to be quarantined and handled in special process
conflict_get_latest_df = get_latest_records_with_pk_confict_detection_flag(df, primary_key_columns, order_by_columns)
conflict_get_latest_df.show(truncate=True)
>> +---+-------------------+------+-----------------+
>> | pk| change_ts| attr|__has_pk_conflict|
>> +---+-------------------+------+-----------------+
>> | 1|2000-01-01 02:00:00| 1002| false|
>> | 2|2000-01-01 00:00:00| 2001| false|
>> | 3|2000-01-01 02:00:00|3002#2| true|
>> | 3|2000-01-01 02:00:00|3002#1| true|
>> +---+-------------------+------+-----------------+
Given list of possible columns, constructs the lists of all possible combinations of composite primary keys, and executes concurrently to determine if given set of columns is a valid primary key. Uses minimum possible amount of queries against spark by skipping validation paths that are based on already validated primary key combinations.
import bdq
import pyspark.sql.functions as F
df = spark.range(0, 100) \
.withColumn('type', F.expr('(id / 10)::int + 1')) \
.withColumn('reminder', F.expr('id % 10')) \
.withColumn('static', F.lit('A')) \
.withColumn('round_robin', F.expr('id % 2'))
display(df)
all_combinations = list(bdq.get_column_names_combinations(df.columns))
bdq.validate_primary_key_candidate_combinations(df, all_combinations, max_workers=10, verbose=True)
>> [('id',), ('type', 'reminder')]
import bdq
import time
import pytest
#create graph
graph = bdq.DAG()
#define nodes
@graph.node()
def a():
time.sleep(2)
return 5
@graph.node()
def b():
time.sleep(3)
return "beeep"
#nodes can depend on other nodes
@graph.node(depends_on=[b])
def c():
time.sleep(5)
return 8
#any amount of dependencies is allowed
@graph.node(depends_on=[b, c, a])
def d():
time.sleep(7)
#beep as many times as absolute difference between 'c' and 'a'
return "g man say: " + b.result * abs(c.result - a.result)
@graph.node(depends_on=[a])
def e():
time.sleep(3)
#you can refer to parent nodes, to get information about their results
raise ValueError(f"omg, crash! {a.result}")
@graph.node(depends_on=[e])
def f():
print("this will never execute")
return "this will never execute"
@graph.node(depends_on=[a])
def g():
time.sleep(3)
#we do not want to execute children nodes anymore
#return graph.BREAK this will cause that all children/successors in graph will be skipped without errror
return graph.BREAK
@graph.node(depends_on=[g])
def i():
print("this will never execute too")
return "this will never execute too"
# if you have ipydagred3 installed, you can see in real time state changes of each of the nodes
display(graph.visualize())
#execute DAG
graph.execute(max_workers=10)
#iterate over results
print("Iterate over results...")
for node in graph.nodes:
print(node)
>> Waiting for all tasks to finish...
>> starting: Node(<function a at 0x7f50f4b9ee50>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> starting: Node(<function b at 0x7f50f4b53040>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> starting: Node(<function e at 0x7f50f4b53b80>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> starting: Node(<function g at 0x7f50f4b535e0>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(<function a at 0x7f50f4b9ee50>: {'state': 'SUCCESS', 'result': 5, 'exception': None, 'completed': True} ) (still running: 3)
>> starting: Node(<function c at 0x7f50f4b534c0>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(<function b at 0x7f50f4b53040>: {'state': 'SUCCESS', 'result': 'beeep', 'exception': None, 'completed': True} ) (still running: 3)
>> error: Node(<function e at 0x7f50f4b53b80>: {'state': 'ERROR', 'result': None, 'exception': ValueError('omg, crash! 5'), 'completed': True} ): omg, crash! 5 (still running: 2)
>> finished: Node(<function g at 0x7f50f4b535e0>: {'state': 'SKIPPED', 'result': <threading.Event object at 0x7f50ec16c9d0>, 'exception': None, 'completed': True} ) (still running: 1)
>> starting: Node(<function d at 0x7f50f4b53a60>: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(<function c at 0x7f50f4b534c0>: {'state': 'SUCCESS', 'result': 8, 'exception': None, 'completed': True} ) (still running: 1)
>> finished: Node(<function d at 0x7f50f4b53a60>: {'state': 'SUCCESS', 'result': 'g man say: beeepbeeepbeeep', 'exception': None, 'completed': True} ) (still running: 0)
>> All tasks finished, shutting down
>>
>> Iterate over results...
>> Node(<function a at 0x7f50f4b9ee50>: {'state': 'SUCCESS', 'result': 5, 'exception': None, 'completed': True} )
>> Node(<function b at 0x7f50f4b53040>: {'state': 'SUCCESS', 'result': 'beeep', 'exception': None, 'completed': True} )
>> Node(<function c at 0x7f50f4b534c0>: {'state': 'SUCCESS', 'result': 8, 'exception': None, 'completed': True} )
>> Node(<function d at 0x7f50f4b53a60>: {'state': 'SUCCESS', 'result': 'g man say: beeepbeeepbeeep', 'exception': None, 'completed': True} )
>> Node(<function e at 0x7f50f4b53b80>: {'state': 'ERROR', 'result': None, 'exception': ValueError('omg, crash! 5'), 'completed': True} )
>> Node(<function f at 0x7f50f4b530d0>: {'state': 'SKIPPED', 'result': None, 'exception': None, 'completed': False} )
>> Node(<function g at 0x7f50f4b535e0>: {'state': 'SKIPPED', 'result': <threading.Event object at 0x7f50ec16c9d0>, 'exception': None, 'completed': True} )
>> Node(<function i at 0x7f50f4b53280>: {'state': 'SKIPPED', 'result': None, 'exception': None, 'completed': False} )
using DAG component to handle parallel execution
import bdq
from bdq import spark, table
ppn = bdq.SparkPipeline("sample")
# returns dataframe, and creates spark view 'raw_data_single_source'
@ppn.step_spark_temp_view()
def raw_data_single_source(step):
return spark.range(1, 10)
# returns dataframe, and creates spark view 'raw_nice_name'
@ppn.step_spark_temp_view(outputs="raw_nice_name")
def raw_data_single_source_with_custom_name(step):
return spark.range(100, 110)
# returns two dataframes, and creates two spark views 'raw_data1', 'raw_data2'
@ppn.step_spark_temp_view(outputs=["raw_data1", "raw_data2"])
def raw_data_multi_source(step):
df1 = spark.range(1000, 2000)
df2 = spark.range(2000, 3000)
return [df1, df2]
# waits for raw data sources to finish, and combines the data into one unioned view `combine_data`
# note that dependencies are python functions or names of outputs
@ppn.step_spark_temp_view(depends_on=[raw_data_single_source, raw_data_single_source_with_custom_name, 'raw_data1', 'raw_data2'])
def combine_data(step):
df = table('raw_data_single_source') \
.union(table('raw_nice_name')) \
.union(table('raw_data1')) \
.union(table('raw_data2'))
return df
# splits the combined_data into 'odd' and 'even' views
@ppn.step_spark_temp_view(depends_on=combine_data, outputs=['odd', 'even'])
def split_data(step):
df_odd = table('combine_data').filter('id % 2 == 1')
df_even = table('combine_data').filter('id % 2 == 0')
return [ df_odd, df_even ]
# executes pipeline using concurrent threads, one per each step, following the dependency DAG
# pipeline is a normal python callable object, as if it was a function, it returns a list of all steps
pipeline_results = ppn(max_concurrent_steps=10)
print('pipeline results:')
print(pipeline_results)
#show some final values
print('even numbers:')
print(table('even').limit(10).collect())
print('odd numbers:')
print(table('odd').limit(10).collect())
#get skipped steps
assert list(ppn.skipped_steps) == []
#get errored steps (you would need to 'adjust' code of on of the steps to make it fail to see something here)
assert list(ppn.error_steps) == []
#get successfull steps
assert set(ppn.success_steps.values()) == set([
raw_data_single_source_with_custom_name,
raw_data_multi_source,
split_data,
combine_data,
raw_data_single_source
])
>> Waiting for all tasks to finish...
>> starting: Node(raw_data_single_source: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> starting: Node(raw_data_single_source_with_custom_name: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> starting: Node(raw_data_multi_source: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(raw_data_single_source: {'state': 'SUCCESS', 'result': [DataFrame[id: bigint]], 'exception': None, 'completed': True} ) (still running: 2)
>> finished: Node(raw_data_single_source_with_custom_name: {'state': 'SUCCESS', 'result': [DataFrame[id: bigint]], 'exception': None, 'completed': True} ) (still running: 1)
>> starting: Node(combine_data: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(raw_data_multi_source: {'state': 'SUCCESS', 'result': [DataFrame[id: bigint], DataFrame[id: bigint]], 'exception': None, 'completed': True} ) (still running: 1)
>> starting: Node(split_data: {'state': 'RUNNING', 'result': None, 'exception': None, 'completed': False} )
>> finished: Node(combine_data: {'state': 'SUCCESS', 'result': [DataFrame[id: bigint]], 'exception': None, 'completed': True} ) (still running: 1)
>> finished: Node(split_data: {'state': 'SUCCESS', 'result': [DataFrame[id: bigint], DataFrame[id: bigint]], 'exception': None, 'completed': True} ) (still running: 0)
>> All tasks finished, shutting down
>> pipeline results:
>> [raw_data_single_source, raw_data_single_source_with_custom_name, raw_data_multi_source, combine_data, split_data]
>> even numbers:
>> [Row(id=2), Row(id=4), Row(id=6), Row(id=8), Row(id=100), Row(id=102), Row(id=104), Row(id=106), Row(id=108), Row(id=1000)]
>> odd numbers:
>> [Row(id=1), Row(id=3), Row(id=5), Row(id=7), Row(id=9), Row(id=101), Row(id=103), Row(id=105), Row(id=107), Row(id=109)]
pipeline steps are rerunable as any ordinary function:
# to rerun given step, just execute it as if it was a pure function
# return is alaways a list of dataframs that given @ppn.step returns
# note: the spark view 'raw_data_single_source' will be updated when this function finishes (as per defintion in @ppn.step above)
raw_data_single_source()
When running code using pyspark, spark ui gets very crowded. SparkUILogger
context manager and decorator assings human readable names to spark stages.
from bdq import SparkUILogger
# usage of decorators
# spark ui stages will have description of 'xyz'
@SparkUILogger.tag(desc='xyz')
def some_function2(number):
return spark.range(number).count()
# spark ui stages will have description of 'alpha_function2' (automatically derived from function name)
@SparkUILogger.tag
def alpha_function2(number):
# 1st and 2nd some_function2() calls should be visible in spark ui as 'xyz'
# the 3rd part of result should be visible as 'alpha_function2
return some_function2(number*2) + some_function2(number*3) + spark.range(number/10).count()
# usage of context managers
# will be visible in spark ui as 'first-count'
with SparkUILogger('first-count'):
spark.range(10).count()
# will be visible in spark ui as '2nd-count'
with SparkUILogger('2nd-count'):
spark.range(20).count()
some_function2(1000)
alpha_function2(2000)
FAQs
Big Data Quality, a set of tools/function that help you every day assert quality of datasets you have processed or ingested
We found that bdq demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Research
Security News
A malicious npm package targets Solana developers, rerouting funds in 2% of transactions to a hardcoded address.
Security News
Research
Socket researchers have discovered malicious npm packages targeting crypto developers, stealing credentials and wallet data using spyware delivered through typosquats of popular cryptographic libraries.
Security News
Socket's package search now displays weekly downloads for npm packages, helping developers quickly assess popularity and make more informed decisions.