Security News
Fluent Assertions Faces Backlash After Abandoning Open Source Licensing
Fluent Assertions is facing backlash after dropping the Apache license for a commercial model, leaving users blindsided and questioning contributor rights.
The perfmetrics package provides a simple way to add software performance metrics to Python libraries and applications. Use perfmetrics to find the true bottlenecks in a production application.
The perfmetrics package is a client of the Statsd daemon by Etsy, which is in turn a client of Graphite (specifically, the Carbon daemon). Because the perfmetrics package sends UDP packets to Statsd, perfmetrics adds no I/O delays to applications and little CPU overhead. It can work equally well in threaded (synchronous) or event-driven (asynchronous) software.
Complete documentation is hosted at https://perfmetrics.readthedocs.io
.. image:: https://img.shields.io/pypi/v/perfmetrics.svg :target: https://pypi.org/project/perfmetrics/ :alt: Latest release
.. image:: https://img.shields.io/pypi/pyversions/perfmetrics.svg :target: https://pypi.org/project/perfmetrics/ :alt: Supported Python versions
.. image:: https://github.com/zodb/perfmetrics/workflows/tests/badge.svg :target: https://github.com/zodb/perfmetrics/actions?query=workflow%3Atests :alt: CI Build Status
.. image:: https://coveralls.io/repos/github/zodb/perfmetrics/badge.svg :target: https://coveralls.io/github/zodb/perfmetrics :alt: Code Coverage
.. image:: https://readthedocs.org/projects/perfmetrics/badge/?version=latest :target: https://perfmetrics.readthedocs.io/en/latest/?badge=latest :alt: Documentation Status
Use the @metric
and @metricmethod
decorators to wrap functions
and methods that should send timing and call statistics to Statsd.
Add the decorators to any function or method that could be a bottleneck,
including library functions.
.. caution::
These decorators are generic and cause the actual function
signature to be lost, replaced with *args, **kwargs
. This can
break certain types of introspection, including zope.interface validation <https://github.com/zodb/perfmetrics/issues/15>
_. As a
workaround, setting the environment variable
PERFMETRICS_DISABLE_DECORATOR
before importing perfmetrics or
code that uses it will cause @perfmetrics.metric
, @perfmetrics.metricmethod
,
@perfmetrics.Metric(...)
and @perfmetrics.MetricMod(...)
to
return the original function unchanged.
Sample::
from perfmetrics import metric
from perfmetrics import metricmethod
@metric
def myfunction():
"""Do something that might be expensive"""
class MyClass(object):
@metricmethod
def mymethod(self):
"""Do some other possibly expensive thing"""
Next, tell perfmetrics how to connect to Statsd. (Until you do, the decorators have no effect.) Ideally, either your application should read the Statsd URI from a configuration file at startup time, or you should set the STATSD_URI environment variable. The example below uses a hard-coded URI::
from perfmetrics import set_statsd_client
set_statsd_client('statsd://localhost:8125')
for i in xrange(1000):
myfunction()
MyClass().mymethod()
If you run that code, it will fire 2000 UDP packets at port 8125. However, unless you have already installed Graphite and Statsd, all of those packets will be ignored and dropped. Dropping is a good thing: you don't want your production application to fail or slow down just because your performance monitoring system is stopped or not working.
Install Graphite and Statsd to receive and graph the metrics. One good way
to install them is the graphite_buildout example
_ at github, which
installs Graphite and Statsd in a custom location without root access.
.. _graphite_buildout example
: https://github.com/hathawsh/graphite_buildout
If you have a Pyramid app, you can set the statsd_uri
for each
request by including perfmetrics in your configuration::
config = Configuration(...)
config.include('perfmetrics')
Also add a statsd_uri
setting such as statsd://localhost:8125
.
Once configured, the perfmetrics tween will set up a Statsd client for
the duration of each request. This is especially useful if you run
multiple apps in one Python interpreter and you want a different
statsd_uri
for each app.
Similar functionality exists for WSGI apps. Add the app to your Paste Deploy pipeline::
[statsd]
use = egg:perfmetrics#statsd
statsd_uri = statsd://localhost:8125
[pipeline:main]
pipeline =
statsd
egg:myapp#myentrypoint
While most programs send metrics from any thread to a single global
Statsd server, some programs need to use a different Statsd server
for each thread. If you only need a global Statsd server, use the
set_statsd_client
function at application startup. If you need
to use a different Statsd server for each thread, use the
statsd_client_stack
object in each thread. Use the
push
, pop
, and clear
methods.
Graphite stores each metric as a time series with multiple resolutions. The sample graphite_buildout stores 10 second resolution for 48 hours, 1 hour resolution for 31 days, and 1 day resolution for 5 years. To produce a coarse grained value from a fine grained value, Graphite computes the mean value (average) for each time span.
Because Graphite computes mean values implicitly, the most sensible way to treat counters in Graphite is as a "hits per second" value. That way, a graph can produce correct results no matter which resolution level it uses.
Treating counters as hits per second has unfortunate consequences, however. If some metric sees a 1000 hit spike in one second, then falls to zero for at least 9 seconds, the Graphite chart for that metric will show a spike of 100, not 1000, since Graphite receives metrics every 10 seconds and the spike looks to Graphite like 100 hits per second over a 10 second period.
If you want your graph to show 1000 hits rather than 100 hits per second,
apply the Graphite hitcount()
function, using a resolution of
10 seconds or more. The hitcount function converts per-second
values to approximate raw hit counts. Be sure
to provide a resolution value large enough to be represented by at least
one pixel width on the resulting graph, otherwise Graphite will compute
averages of hit counts and produce a confusing graph.
It usually makes sense to treat null values in Graphite as zero, though that is not the default; by default, Graphite draws nothing for null values. You can turn on that option for each graph.
NOTE: This will be the last major release to support legacy versions of Python such as 2.7 and 3.6. Some such legacy versions may not have binary wheels published for this release.
Add support for Python 3.10.
Drop support for Python 3.5.
Add aarch64 binary wheels.
issue 26 <https://github.com/zodb/perfmetrics/issues/26>
_.FakeStatsDClient
for testing is now always true whether or
not any observations have been seen, like the normal clients. See
issue <https://github.com/zodb/perfmetrics/issues/23>
_.StatsD sets <https://github.com/statsd/statsd/blob/master/docs/metric_types.md#sets>
,
counters of unique events. See PR 30 <https://github.com/zodb/perfmetrics/pull/30>
.Drop support for EOL Python 2.6, 3.2, 3.3 and 3.4.
Add support for Python 3.5, 3.6, and 3.7.
Compile the performance-sensitive parts with Cython, leading to a 10-30% speed improvement. See https://github.com/zodb/perfmetrics/issues/17.
Caution: Metric names are enforced to be native strings (as a result
of Cython compilation); they've always had to be ASCII-only but
previously Unicode was allowed on Python 2. This is usually
automatically the case when used as a decorator. On Python 2 using
from __future__ import unicode_literals
can cause problems
(raising TypeError) when manually constructing Metric
objects. A
quick workaround is to set the environment variable
PERFMETRICS_PURE_PYTHON
before importing perfmetrics.
Make decorated functions and methods configurable at runtime, not just compile time. See https://github.com/zodb/perfmetrics/issues/11.
Include support for testing applications instrumented with
perfmetrics in perfmetrics.testing
. This was previously released
externally as nti.fakestatsd
. See https://github.com/zodb/perfmetrics/issues/9.
Read the PERFMETRICS_DISABLE_DECORATOR
environment variable when
perfmetrics
is imported, and if it is set, make the decorators @metric
,
@metricmethod
, @Metric(...)
and @MetricMod(...)
return
the function unchanged. This can be helpful for certain kinds of
introspection tests. See https://github.com/zodb/perfmetrics/issues/15
Added the @MetricMod
decorator, which changes the name of
metrics in a given context. For example, @MetricMod('xyz.%s')
adds a prefix.
Removed the "gauge suffix" feature. It was unnecessarily confusing.
Timing metrics produced by @metric
, @metricmethod
, and
@Metric
now have a ".t" suffix by default to avoid naming
conflicts.
STATSD_URI
environment variable.Metric
can now be used as either a decorator or a context
manager.
Made the signature of StatsdClient more like James Socol's StatsClient.
FAQs
Send performance metrics about Python code to Statsd
We found that perfmetrics demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 2 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
Fluent Assertions is facing backlash after dropping the Apache license for a commercial model, leaving users blindsided and questioning contributor rights.
Research
Security News
Socket researchers uncover the risks of a malicious Python package targeting Discord developers.
Security News
The UK is proposing a bold ban on ransomware payments by public entities to disrupt cybercrime, protect critical services, and lead global cybersecurity efforts.