
Security News
Crates.io Implements Trusted Publishing Support
Crates.io adds Trusted Publishing support, enabling secure GitHub Actions-based crate releases without long-lived API tokens.
py_zipkin provides a context manager/decorator along with some utilities to facilitate the usage of Zipkin in Python applications.
pip install py_zipkin
py_zipkin requires a transport_handler
object that handles logging zipkin
messages to a central logging service such as kafka or scribe.
py_zipkin.zipkin.zipkin_span
is the main tool for starting zipkin traces or
logging spans inside an ongoing trace. zipkin_span can be used as a context
manager or a decorator.
from py_zipkin.zipkin import zipkin_span
def some_function(a, b):
with zipkin_span(
service_name='my_service',
span_name='my_span_name',
transport_handler=some_handler,
port=42,
sample_rate=0.05, # Value between 0.0 and 100.0
):
do_stuff(a, b)
The difference between this and Usage #1 is that the zipkin_attrs are calculated separately and passed in, thus negating the need of the sample_rate param.
# Define a pyramid tween
def tween(request):
zipkin_attrs = some_zipkin_attr_creator(request)
with zipkin_span(
service_name='my_service',
span_name='my_span_name',
zipkin_attrs=zipkin_attrs,
transport_handler=some_handler,
port=22,
) as zipkin_context:
response = handler(request)
zipkin_context.update_binary_annotations(
some_binary_annotations)
return response
This can be also be used inside itself to produce continuously nested spans.
@zipkin_span(service_name='my_service', span_name='some_function')
def some_function(a, b):
return do_stuff(a, b)
zipkin_span.update_binary_annotations()
can be used inside a zipkin trace
to add to the existing set of binary annotations.
def some_function(a, b):
with zipkin_span(
service_name='my_service',
span_name='some_function',
transport_handler=some_handler,
port=42,
sample_rate=0.05,
) as zipkin_context:
result = do_stuff(a, b)
zipkin_context.update_binary_annotations({'result': result})
zipkin_span.add_sa_binary_annotation()
can be used to add a binary annotation
to the current span with the key 'sa'. This function allows the user to specify the
destination address of the service being called (useful if the destination doesn't
support zipkin). See http://zipkin.io/pages/data_model.html for more information on the
'sa' binary annotation.
NOTE: the V2 span format only support 1 "sa" endpoint (represented by remoteEndpoint) so
add_sa_binary_annotation
now raisesValueError
if you try to set multiple "sa" annotations for the same span.
def some_function():
with zipkin_span(
service_name='my_service',
span_name='some_function',
transport_handler=some_handler,
port=42,
sample_rate=0.05,
) as zipkin_context:
make_call_to_non_instrumented_service()
zipkin_context.add_sa_binary_annotation(
port=123,
service_name='non_instrumented_service',
host='12.34.56.78',
)
create_http_headers_for_new_span()
creates a set of HTTP headers that can be forwarded
in a request to another service.
headers = {}
headers.update(create_http_headers_for_new_span())
http_client.get(
path='some_url',
headers=headers,
)
py_zipkin (for the moment) thrift-encodes spans. The actual transport layer is pluggable, though.
The recommended way to implement a new transport handler is to subclass
py_zipkin.transport.BaseTransportHandler
and implement the send
and
get_max_payload_bytes
methods.
send
receives an already encoded thrift list as argument.
get_max_payload_bytes
should return the maximum payload size supported by your
transport, or None
if you can send arbitrarily big messages.
The simplest way to get spans to the collector is via HTTP POST. Here's an
example of a simple HTTP transport using the requests
library. This assumes
your Zipkin collector is running at localhost:9411.
NOTE: older versions of py_zipkin suggested implementing the transport handler as a function with a single argument. That's still supported and should work with the current py_zipkin version, but it's deprecated.
import requests
from py_zipkin.transport import BaseTransportHandler
class HttpTransport(BaseTransportHandler):
def get_max_payload_bytes(self):
return None
def send(self, encoded_span):
# The collector expects a thrift-encoded list of spans.
requests.post(
'http://localhost:9411/api/v1/spans',
data=encoded_span,
headers={'Content-Type': 'application/x-thrift'},
)
If you have the ability to send spans over Kafka (more like what you might do in production), you'd do something like the following, using the kafka-python package:
from kafka import SimpleProducer, KafkaClient
from py_zipkin.transport import BaseTransportHandler
class KafkaTransport(BaseTransportHandler):
def get_max_payload_bytes(self):
# By default Kafka rejects messages bigger than 1000012 bytes.
return 1000012
def send(self, message):
kafka_client = KafkaClient('{}:{}'.format('localhost', 9092))
producer = SimpleProducer(kafka_client)
producer.send_messages('kafka_topic_name', message)
If you want to use py_zipkin in a cooperative multithreading environment,
e.g. asyncio, you need to explicitly pass an instance of py_zipkin.storage.Stack
as parameter context_stack
for zipkin_span
and create_http_headers_for_new_span
.
By default, py_zipkin uses a thread local storage for the attributes, which is
defined in py_zipkin.storage.ThreadLocalStack
.
Additionally, you'll also need to explicitly pass an instance of
py_zipkin.storage.SpanStorage
as parameter span_storage
to zipkin_span
.
from py_zipkin.zipkin import zipkin_span
from py_zipkin.storage import Stack
from py_zipkin.storage import SpanStorage
def my_function():
context_stack = Stack()
span_storage = SpanStorage()
await my_function(context_stack, span_storage)
async def my_function(context_stack, span_storage):
with zipkin_span(
service_name='my_service',
span_name='some_function',
transport_handler=some_handler,
port=42,
sample_rate=0.05,
context_stack=context_stack,
span_storage=span_storage,
):
result = do_stuff(a, b)
"Firehose mode" records 100% of the spans, regardless of
sampling rate. This is useful if you want to treat these spans
differently, e.g. send them to a different backend that has limited
retention. It works in tandem with normal operation, however there may
be additional overhead. In order to use this, you add a
firehose_handler
just like you add a transport_handler
.
This feature should be considered experimental and may be removed at any time without warning. If you do use this, be sure to send asynchronously to avoid excess overhead for every request.
Copyright (c) 2018, Yelp, Inc. All Rights reserved. Apache v2
zipkin_logger.debug()
interface.py_zipkin.stack
was renamed as py_zipkin.storage
. If you were
importing this module, you'll need to update your code.ThreadLocalStack()
if passed as
None
py_zipkin.thread_local
is now deprecated. Instead use
py_zipkin.stack.ThreadLocalStack()
update_binary_annotations
would fail for a child
span in a trace that is not being sampledupdate_binary_annotations
for both root and non-root spanszipkin_span
to report timestamp/duration.
Changes API of zipkin_span
, but defaults back to existing behavior.zipkin_span
entrypoint should be backwards compatiblestart
and stop
functions as friendlier versions of the
enter and exit functions.FAQs
Library for using Zipkin in Python.
We found that py-zipkin demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 3 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
Crates.io adds Trusted Publishing support, enabling secure GitHub Actions-based crate releases without long-lived API tokens.
Research
/Security News
Undocumented protestware found in 28 npm packages disrupts UI for Russian-language users visiting Russian and Belarusian domains.
Research
/Security News
North Korean threat actors deploy 67 malicious npm packages using the newly discovered XORIndex malware loader.