Security News
38% of CISOs Fear They’re Not Moving Fast Enough on AI
CISOs are racing to adopt AI for cybersecurity, but hurdles in budgets and governance may leave some falling behind in the fight against cyber threats.
Module and command-line tool that wraps around hashlib and zlib to facilitate generating checksums / hashes of files and directories.
filehash
.. image:: https://img.shields.io/pypi/v/filehash.svg :target: https://pypi.python.org/pypi/filehash
.. image:: https://img.shields.io/travis/leonidessaguisagjr/filehash.svg :target: https://travis-ci.org/leonidessaguisagjr/filehash
Python module to facilitate calculating the checksum or hash of a file. Tested against Python 2.7.x, Python 3.6.x, Python 3.7.x, Python 3.8.x, Python 3.9.x, Python 3.10.x, PyPy 2.7.x and PyPy3 3.7.x. Currently supports Adler-32 <https://en.wikipedia.org/wiki/Adler-32>
, BLAKE2b, BLAKE2s <https://en.wikipedia.org/wiki/BLAKE_(hash_function)#BLAKE2>
, CRC32 <https://en.wikipedia.org/wiki/Cyclic_redundancy_check>
, MD5 <https://en.wikipedia.org/wiki/MD5>
, SHA-1 <https://en.wikipedia.org/wiki/SHA-1>
, SHA-224, SHA-256, SHA-384 and SHA-512 <https://en.wikipedia.org/wiki/SHA-2>
.
(Note: BLAKE2b and BLAKE2s are only supported on Python 3.6.x and later.)
FileHash
classThe FileHash
class wraps around the hashlib
(provides hashing for MD5, SHA-1, SHA-224, SHA-256, SHA-384 and SHA-512) and zlib
(provides checksums for Adler-32 and CRC32) modules and contains the following methods:
hash_file(filename)
- Calculate the file hash for a single file. Returns a string with the hex digest.hash_files(filename)
- Calculate the file hash for multiple files. Returns a list of tuples where each tuple contains the filename and the calculated hash.hash_dir(path, pattern='*')
- Calculate the file hashes for an entire directory. Returns a list of tuples where each tuple contains the filename and the calculated hash.cathash_files(filenames)
- Calculate a single hash for multiple files. Files are sorted by their individual hash values and then traversed in that order to generate a combined hash value. Returns a string with the hex digest.cathash_dir(path, pattern='*')
- Calculate a single hash for an entire directory of files. Files are sorted by their individual hash values and then traversed in that order to generate a combined hash value. Returns a string with the hex digest.verify_sfv(sfv_filename)
- Reads the specified SFV (Simple File Verification) file and calculates the CRC32 checksum for the files listed, comparing the calculated CRC32 checksums against the specified expected checksums. Returns a list of tuples where each tuple contains the filename and a boolean value indicating if the calculated CRC32 checksum matches the expected CRC32 checksum. To find out more about SFV files, see the Simple file verification entry in Wikipedia <https://en.wikipedia.org/wiki/Simple_file_verification>
_.verify_checksums(checksum_filename)
- Reads the specified file and calculates the hashes for the files listed, comparing the calculated hashes against the specified expected hashes. Returns a list of tuples where each tuple contains the filename and a boolean value indicating if the calculated hash matches the expected hash.For the checksum file, the file is expected to be a plain text file where each line has an entry formatted as follows::
{hash}[SPACE][ASTERISK]{filename}
This format is the format used by programs such as the sha1sum <https://en.wikipedia.org/wiki/Sha1sum>
_ family of tools for generating checksum files. Here is an example generated by sha1sum
::
f7ef3b7afaf1518032da1b832436ef3bbfd4e6f0 *lorem_ipsum.txt 03da86258449317e8834a54cf8c4d5b41e7c7128 *lorem_ipsum.zip
The FileHash
constructor has two optional arguments:
hash_algorithm='sha256'
- Specifies the hashing algorithm to use. See filehash.SUPPORTED_ALGORITHMS
for the list of supported hash / checksum algorithms. Defaults to SHA256.chunk_size=4096
- Integer specifying the chunk size to use (in bytes) when reading the file. This comes in useful when processing very large files to avoid having to read the entire file into memory all at once. Default chunk size is 4096 bytes.The library can be used as follows::
import os from filehash import FileHash md5hasher = FileHash('md5') md5hasher.hash_file("./testdata/lorem_ipsum.txt") '72f5d9e3a5fa2f2e591487ae02489388' sha1hasher = FileHash('sha1') sha1hasher.hash_dir("./testdata", "*.zip") [FileHashResult(filename='lorem_ipsum.zip', hash='03da86258449317e8834a54cf8c4d5b41e7c7128')] sha512hasher = FileHash('sha512') os.chdir("./testdata") sha512hasher.verify_checksums("./hashes.sha512") [VerifyHashResult(filename='lorem_ipsum.txt', hashes_match=True), VerifyHashResult(filename='lorem_ipsum.zip', hashes_match=True)] crc32hasher = FileHash('crc32') crc32hasher.verify_sfv("./lorem_ipsum.sfv") [VerifyHashResult(filename='lorem_ipsum.txt', hashes_match=True), VerifyHashResult(filename='lorem_ipsum.zip', hashes_match=True)]
chkfilehash
command line toolA command-line tool called chkfilehash
is also included with the filehash
package. Here is an example of how the tool can be used::
$ chkfilehash -a sha512 -c hashes.sha512 lorem_ipsum.txt: OK lorem_ipsum.zip: OK $ chkfilehash -a crc32 lorem_ipsum.zip 7425D3BE *lorem_ipsum.zip $
Run the tool without any parameters or with the -h
/ --help
switch to get a usage screen.
This is released under an MIT license. See the LICENSE
file in this repository for more information.
FAQs
Module and command-line tool that wraps around hashlib and zlib to facilitate generating checksums / hashes of files and directories.
We found that filehash demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 1 open source maintainer collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
CISOs are racing to adopt AI for cybersecurity, but hurdles in budgets and governance may leave some falling behind in the fight against cyber threats.
Research
Security News
Socket researchers uncovered a backdoored typosquat of BoltDB in the Go ecosystem, exploiting Go Module Proxy caching to persist undetected for years.
Security News
Company News
Socket is joining TC54 to help develop standards for software supply chain security, contributing to the evolution of SBOMs, CycloneDX, and Package URL specifications.