Note: most of this file and library was written in 2011 and not modified since. Some of the advice may be slightly dated (e.g. talking of GCC 4.4.3 as "recent"). The code should still work properly.
Sphlib is a set of implementations of various hash functions, both in C
and in Java. The C code is meant to be easily imported into other
projects, in particular embedded systems. The Java code implements an
API somewhat similar to that of java.security.MessageDigest.
The C source code also provides two standalone tools:
sphspeed: performs speed tests on various hash functionssphsum: computes and verifies checksums over files
IMPORTANT NOTE: for users of the previous version (sphlib-2.1):
BLAKE, Groestl, JH, Keccak and Skein have been updated, to match the "tweaked" specifications published for the third round of the SHA-3 competition. Thus, these function now return distinct values from what they were producing previously. Also, for Skein with a 224-bit or 256-bit output, the size of the context structure has changed, so calling code must be recompiled as well.
Licensing is specified in the LICENSE file. This is an MIT-like, BSD-like open-source license. Basically, we will get the fame but not the blame. If you reuse our code in your own projects, and distribute the result, then you should state that you used our code and that we always disclaimed any kind of warranty, and will continue to do so in the foreseeable future, and beyond. You have no other obligation such as disclosing your own source code. See the LICENSE file for the details in a lawyer-compatible language.
The authors are the "Projet RNRT SAPHIR", which is a research project sponsored by the French government; project members are public and private organizations:
- Cryptolog
- DCSSI
- École Normale Supérieure
- France Telecom
- Gemalto
Projet RNRT SAPHIR was continued into Projet RNRT SAPHIR2, with four new additional members:
- EADS SN
- Sagem Securite
- INRIA
- UVSQ
We use the "Projet RNRT SAPHIR" expression to designate both SAPHIR and SAPHIR2.
All the actual code has been written by:
Thomas Pornin <pornin@bolet.org>
to whom technical questions may be addressed. Note that I do not claim authorship: all writing was done on behalf of the Projet RNRT SAPHIR.
The programming interface for both the C code and the Java code can be
found in the doc/ subdirectory. This documentation is in HTML format
and was generated from the comments in the source code with,
respectively, doxygen and javadoc.
The hash functions have been implemented with regards to their published specification. Whenever possible, the correction of the implementation has been verified with regards to published test vectors. Some functions have several variants; for instance, there are three distinct "Whirlpool" which sphlib implements, under the names "Whirlpool-0", "Whirlpool-1" and "Whirlpool".
For the SHA-3 candidates, sphlib follows the "round 3" specifications, thus including the "tweaks" that some of the candidates added right after round 1 and all also the tweaks that the "finalists" added after round 2. For some of those functions, the officially submitted code and test vectors turned out to be flawed (non conforming to the specification), and corrections were published by their authors; sphlib follows the specification and agrees with those corrected versions.
For two of the second round SHA-3 candidates (Hamsi and SHAvite-3), the most recently published specifications (as of June 18th, 2010) have some flaws which do not alter the function robustness or performance, but still mean that some or all of the published implementations and test vectors are wrong. The respective designers of those functions are aware of those flaws and intend to publish corrections at some point. sphlib anticipates on those corrections and already implements them.
The c/ subdirectory contain the C code. In that directory, there are
two Makefiles and a build shell script. The shell script, named
build.sh, is for Unix-like systems.
sphlib does not feature a "proper" compilation and configuration system such as those customarily found in open-source libraries for Unix systems. This may be corrected in a future version. Right now, I am not utterly convinced that the autoconf-generated scripts are the "way to go". Anyway, sphlib is meant for evaluation, research and import into other projects; a streamlined standalone compilation process is hardly relevant for those usages.
By default, sphlib compiles for "big" architectures, using heavy loop unrolling. This is what provides the best performance on modern PC, workstations, servers, and about any architecture where the level-1 cache for instruction (in the CPU) has size 32 kB or more.
However, sphlib also includes variants optimized for architectures with
small level-1 cache. To use them, arrange for the SPH_SMALL_FOOTPRINT
macro to be defined (to a non-zero integer value) during compilation,
e.g. through the arguments passed to the C compiler by the build script.
These variants have been tested on a MIPS-compatible processor with 8 kB
of level-1 cache, and they offer much better performance than the normal
code on those architectures. In some specific situations, you might want
to use these "small footprint" variants on big computers as well; test
and measure speed if unsure.
If you happen to have a Unix-like system (e.g. Linux), you may simply type:
c/build.sh
which should:
- compile the library
- compile the tools
- compile the unit tests
- run the unit tests
The library and tools may be installed with:
c/build.sh -i
which will install sphspeed and sphsum in /usr/local/bin, libsph.a in
/usr/local/lib, and the header files (all the sph_*.h files) in
/usr/local/include.
The installation directories and the compilation options can be altered at will with appropriate options. Use:
c/build.sh --help
to access the list of options.
build.sh is only for Unix-like systems such as Linux. This script has
not been thoroughly tested, is very crude, and has only limited
autodetection capabilities. If you are after getting the maximum hashing
speed, or if you want to use the library from a shared object, you will
probably have to specify other compile options. Use --with-cflags to
change the compilation options. For instance:
c/build.sh --with-cflags="-W -Wall -O1 -fPIC -mtune=athlon64"
This selects options for position-independant code, i.e. suitable for a
shared object, and tuned for maximum performance on Ahtlon64-type
processors. It has been noticed that -O1 provides better performance
than -O2 with recent versions (4.4.3) of GCC, although -O2 yields
better code for some of the hash functions.
A realistic example of cross-compilation for a MIPS-compatible architecture would look like this:
c/build.sh --with-cc=mipsel-linux-uclibc-gcc \
--with-clags="-W -Wall -O1 -DSPH_SMALL_FOOTPRINT"
which selects an alternate C compiler, and also defines the
SPH_SMALL_FOOTPRINT macro to use the "small footprint" variants which
offer much better performance on architectures with low L1 cache.
build.sh is not mandatory; you may edit and use the Makefile.unix
file directly.
The sphsum binary can be used to hash files in a way similar to what
the md5sum Linux tool does. The first argument of sphsum must be the
name of a hash function; matching is not case sensitive. Recognized
names are:
| name | function |
|---|---|
haval128_3 |
HAVAL, 128-bit output, 3 passes |
haval128_4 |
HAVAL, 128-bit output, 4 passes |
haval128_5 |
HAVAL, 128-bit output, 5 passes |
haval160_3 |
HAVAL, 160-bit output, 3 passes |
haval160_4 |
HAVAL, 160-bit output, 4 passes |
haval160_5 |
HAVAL, 160-bit output, 5 passes |
haval192_3 |
HAVAL, 192-bit output, 3 passes |
haval192_4 |
HAVAL, 192-bit output, 4 passes |
haval192_5 |
HAVAL, 192-bit output, 5 passes |
haval224_3 |
HAVAL, 224-bit output, 3 passes |
haval224_4 |
HAVAL, 224-bit output, 4 passes |
haval224_5 |
HAVAL, 224-bit output, 5 passes |
haval256_3 |
HAVAL, 256-bit output, 3 passes |
haval256_4 |
HAVAL, 256-bit output, 4 passes |
haval256_5 |
HAVAL, 256-bit output, 5 passes |
md2 |
MD2 |
md4 |
MD4 |
md5 |
MD5 |
panama |
Panama |
radiogatun32 |
RadioGatun[32] |
radiogatun64 |
RadioGatun[64] |
ripemd |
RIPEMD (original function) |
ripemd128 |
RIPEMD-128 (revised function, 128-bit output) |
ripemd160 |
RIPEMD-160 (revised function, 160-bit output) |
rmd |
RIPEMD (original function) |
rmd128 |
RIPEMD-128 (revised function, 128-bit output) |
rmd160 |
RIPEMD-160 (revised function, 160-bit output) |
sha0 |
SHA-0 (original SHA, withdrawn) |
sha1 |
SHA-1 |
sha224 |
SHA-224 |
sha256 |
SHA-256 |
sha384 |
SHA-384 |
sha512 |
SHA-512 |
tiger |
Tiger |
tiger2 |
Tiger2 (Tiger with a modified padding) |
whirlpool |
Whirlpool (2003, current version) |
whirlpool0 |
Whirlpool-0 (2000) |
whirlpool1 |
Whirlpool-1 (2001) |
For the implemented "SHA-3 candidates", there are four names for each function, depending on the hash output size in bits. That size is appended to the base name; e.g. "shabal384" means "the Shabal hash function with a 384-bit output". Here are the base names for the implemented SHA-3 candidates:
| base name | function |
|---|---|
blake |
BLAKE |
bmw |
Blue Midnight Wish |
cubehash |
CubeHash |
echo |
ECHO |
fugue |
Fugue |
groestl |
Groestl |
hamsi |
Hamsi |
jh |
JH |
keccak |
Keccak |
luffa |
Luffa |
shabal |
Shabal |
shavite |
SHAvite-3 |
simd |
SIMD |
skein |
Skein |
Alternatively, the sphsum executable file can be named after one of
these functions, in which case the function name needs not be specified.
Hence, if you install sphsum and create a link (either symbolic or
not) to sphsum named md5sum, then you may use that link as a drop-in
replacement for the standard Linux tool md5sum. This function name
recognition process ignores the .exe, sum and sum.exe suffixes.
On Windows systems, you may use the Makefile.win32 file. This is meant
for Visual C 2005 or later (command-line compiler). Open a "Visual C
console" from the start menu (this is a standard text console with the
environment set up for using cl.exe). Type:
nmake /f makefile.win32
which should compile the code, the unit tests and the standalone binaries. There is no library per se, only a collection of object files.
Other C compilers exist for Windows (e.g. MinGW or the cygwin system). They should be able to process sphlib code with no worry; but we provide no build script or makefile for them.
If you wish to include sphlib C code in your own projects, then you must copy the header and source files which implement the functions you want to use. Here are the dependency rules:
-
sph_types.h: always needed; all other files include it. -
Each function or function family has its own header, e.g.
sph_sha2.hfor the SHA-2 family (SHA-224, SHA-256, SHA-384 and SHA-512). Thesph_sha3.hheader includes thesph_sha2.hfile (for SHA-2) and all the header files for the implemented SHA-3 candidates. -
Each function or function family is implemented in one or a few C files. You need to include C files only for the functions that you actually use. Most of the file names are self-explanatory, but please note the following:
-
Some functions indirectly use the
md_helper.cfile. These are MD4, MD5, all RIPEMD*, all SHA-, Tiger, Tiger2 and all Whirpool. Themd_helper.cfile MUST NOT be compiled by itself: it is a helper file which is included by, for instance,md5.c. Just drop it in the same directory. -
Similarly:
- HAVAL (
haval.c) includeshaval_helper.c - ECHO (
echo.c) and SHAvite-3 (shavite.c) includeaes_helper.c - Hamsi (
hamsi.c) includeshamsi_helper.c
- HAVAL (
-
sha2.cis for SHA-224 and SHA-256. sha2big.c is for SHA-384 and SHA-512. -
speed.candhsum.care the main files for, respectively, thesphspeedandsphsumcommand-line utilities. -
utest.c,utest.hand thetest_*.cfiles are used for the unit tests, which verify that the implementations operate properly. They need not be included in your own project. -
sha3nist.candsha3nist.hare a wrapper used to transform SHA-2 or any of the SHA-3 candidates into functions with the API defined by NIST for the SHA-3 competition. You have to modify thesha3nist.hfile to select the actual candidate (only one at a time, this is an artefact of how the NIST API is defined).
-
Most of the "magic" happens in sph_types.h. This is where one may find
such things as inline assembly for faster little/big-endian word access.
The C code tries to detect (through predefined macros) the kind of architecture on which it is supposed to run. This information can be used to speed up some operations, in particular decoding and encoding of 32-bit and 64-bit words. When the current architecture cannot be detected, sphlib uses some generic code which always works but is somewhat slower. The speed gain obtained through architecture specific code can reach +30% on the fastest functions (less on the slower functions).
Most of the C macros which govern that behaviour are boolean flags. To
explicitly enable the feature, define the macro to a non-zero integer
value, e.g. with -DSPH_LITTLE_ENDIAN=1 (for most C compilers, defining
the macro without an explicit content, with -DSPH_LITTLE_ENDIAN, has
the same effect). To explicitly disable the feature, define the macro to
a zero integer value: -DSPH_LITTLE_ENDIAN=0. If a feature is not
explicitly enabled or disabled, then sph_types.h will try to autodetect
its status.
The following flags are defined:
-
SPH_LITTLE_ENDIANWhen non-zero, sphlib assumes that its 32-bit-or-more integer type (respectively 64-bit-or-more integer type) has size exactly 32 bits (respectively 64 bits), and is encoded in RAM with the little-endian convention.
-
SPH_BIG_ENDIANSimilar to
SPH_LITTLE_ENDIAN, but with the big-endian convention. -
SPH_LITTLE_FASTWhen non-zero, little-endian decoding is assumed to be fast: some functions will thus omit caching decoded words in local variables. This is normally implied by
SPH_LITTLE_ENDIAN. -
SPH_BIG_FASTSimilar to
SPH_LITTLE_FAST, but with big-endian convention. -
SPH_UNALIGNEDThe processor tolerates unaligned 32-bit or 64-bit accesses with only a slight timing penalty.
-
SPH_SPARCV9_GCC_32The target architecture is an UltraSPARC-compatible processor, used in 32-bit mode, and the compiler is GCC.
-
SPH_SPARCV9_GCC_64The target architecture is an UltraSPARC-compatible processor, used in 64-bit mode, and the compiler is GCC.
-
SPH_SPARCV9_GCCThe target architecture is an UltraSPARC-compatible processor, used in 32-bit or 64-bit mode, and the compiler is GCC.
-
SPH_I386_GCCThe target architecture is an x86-compatible processor, used in 32-bit mode, and the compiler is GCC.
-
SPH_I386_MSVCThe target architecture is an x86-compatible processor, used in 32-bit mode, and the compiler is Microsoft Visual C.
-
SPH_AMD64_GCCThe target architecture is an x86-compatible processor, used in 64-bit mode, and the compiler is GCC.
-
SPH_AMD64_MSVCThe target architecture is an x86-compatible processor, used in 64-bit mode, and the compiler is Microsoft Visual C.
-
SPH_SMALL_FOOTPRINTWhen non-zero, "small footprint" variants are compiled. The code is less unrolled, resulting in more compact binary code, at the expense of extra indirections. This macro is never auto-detected. You should use it when the target architecture level-1 cache for instructions is strictly smaller than 32 kB.
There are also function-specific "small footprint" flags, which can
be used to enable or disable "small footprint" variants for each
function independently (the function-specific flag takes precedence
over SPH_SMALL_FOOTPRINT when both are defined). These flags are:
| Compilation flag | Relevant hash function |
|---|---|
SPH_SMALL_FOOTPRINT_BLAKE |
BLAKE |
SPH_SMALL_FOOTPRINT_BMW |
Blue Midnight Wish |
SPH_SMALL_FOOTPRINT_CUBEHASH |
CubeHash |
SPH_SMALL_FOOTPRINT_ECHO |
ECHO |
SPH_SMALL_FOOTPRINT_GROESTL |
Groestl |
SPH_SMALL_FOOTPRINT_HAMSI |
Hamsi |
SPH_SMALL_FOOTPRINT_HAVAL |
HAVAL |
SPH_SMALL_FOOTPRINT_JH |
JH |
SPH_SMALL_FOOTPRINT_KECCAK |
Keccak |
SPH_SMALL_FOOTPRINT_SHA2 |
SHA-224, SHA-256, SHA-384 and SHA-512 |
SPH_SMALL_FOOTPRINT_SHAVITE |
SHAvite-3 |
SPH_SMALL_FOOTPRINT_SIMD |
SIMD |
SPH_SMALL_FOOTPRINT_SKEIN |
Skein |
SPH_SMALL_FOOTPRINT_WHIRLPOOL |
Whirlpool |
Another additional macro is SPH_UPTR. This is not a boolean flag; when
defined, it must evaluate to an unsigned integer type which has the same
size as a pointer. When casting a C pointer to SPH_UPTR and back, the
original pointer must be recovered, and it must be possible to determine
the pointer alignment by looking at the least significant bits of its
value when cast to a SPH_UPTR. SPH_UPTR cannot be defined unless
either SPH_LITTLE_ENDIAN or SPH_BIG_ENDIAN is also defined
(explicitly or auto-detected). If unsure, leave undefined; it has no
influence over performance of most of the implemented functions.
The test_types binary (built as part of the unit tests) prints out,
when executed, a synthetic report on what architecture characteristics
were actually used.
Java code is in the java/ directory. Hash function implementations are
located in the fr.cryptohash package; there is one specific class for
each hash function, a common interface called Digest, and some
non-public helper classes.
The fr.cryptohash.test package contains two standalone applications
(classes with a main() method). The TestDigest application runs the
unit tests. The Speed application runs speed tests, with an output
similar to that provided by the sphspeed tool from the C code. Note
that these tests cannot access the CPU usage by the test process;
instead, they use the "wall clock" time. Hence, speed tests should be
performed on an otherwise idle machine.
The Java code should be compatible both with older virtual machines (e.g. Java 1.1) and with J2ME platforms.
IMPORTANT WARNING
It appears that some versions of the Java virtual machine from Sun
(now Oracle) have a bug, in which the code for ECHO is not properly
handled at runtime. To check whether your VM has the bug, run the
fr.cryptohash.test.TestDigest application, preferably with the -server
command-line flag (this is the default on x86_64 but not on i386).
Affected versions include at least 1.6.0_16. However, 1.6.0_19 and 1.6.0_20 seem fine. If unsure then update your JVM to the latest published version.
Some OpenJDK versions are also affected, including 6b16-1.6.1.
Internally, sphlib tended to use the name "sha3" for the 64-bit functions of the SHA-2 family, namely SHA-384 and SHA-512. This is historical. Here, we talk about the SHA-3 contest which was launched in 2008 by NIST, to define the next family of hash functions which will become an american standard, as substitutes for the existing SHA-224, SHA-256, SHA-384 and SHA-512 functions. Many candidate functions have been submitted so far. The competition has reached its second round, in which 14 candidates have been kept. sphlib currently implements those 14 candidates.
For the purposes of this competition, the NIST published a C API. All candidates were asked to provide reference and optimized implementations fitting in that API.
The basic sphlib API is distinct from the NIST API. However, a
compatibility layer has been added to sphlib-1.1. It consists in the
sha3nist.c and sha3nist.h source files. With these files, you may use
some of the sphlib implementations through an API conforming to the NIST
specification. Namely, you may select either the SHA-2 family, or any of
the implemented SHA-3 candidates.
To use that layer, modify sha3nist.h to designate the hash functions
you wish to use. By default, the SHA-224/... functions are used. To
use Shabal instead, replace the following line:
#define SPH_NIST sha
with this:
#define SPH_NIST shabal
and add the sha3nist.c file to the list of C files to compile into your
application. Similarly, use bmw for Blue Midnight Wish, jh for JH,
and so on.
Future versions of sphlib may feature:
-
options for better conditional inclusion (e.g. not compiling RIPEMD if you only want RIPEMD-160)
-
optimized versions for footprint-constrained environments (which should also help platforms with a small L1 cache)
-
a better compilation and installation procedure for the library and standalone tools
-
man pages for the standalone tools
-
a building process for sphlib as a shared library
-
new in sphlib-3.0:
- Updated BLAKE, Groestl, JH, Keccak and Skein to SHA-3 round 3 tweaks
- Fixed suboptimal code in Keccak
- Fixed a data-management bug in Hamsi
-
new in sphlib-2.1:
- Added implementations of CubeHash, Groestl, Hamsi, Keccak and SHAvite-3 (C and Java)
- Added Java implementations for RadioGatun
- Optimized RadioGatun on small architectures and 32-bit x86
- Made "size-generic" Java implementation of Shabal (supports all output sizes multiple of 32, from 32 to 512 bits)
- Added macros for explicit architecture feature activation or deactivation
- Renamed SHABAL -> Shabal, and WHIRLPOOL -> Whirlpool
- Fixed some bugs on exotic architectures
-
new in sphlib-2.0:
- Added implementations of BLAKE, Blue Midnight Wish, ECHO, Fugue, JH, Luffa, SIMD and Skein (C and Java)
- Changed default optimization level to
-O1with GCC - Moved SHA-384 / SHA-512 headers to
sph_sha2.h;sph_sha3.hnow includessph_sha2.hand the header files for all SHA-3 candidates - Renamed implementation file for SHA-384 / SHA-512 (now
sha2big.c) - Added support for signed integer types of at least 32 or 64 bits
- Improved MIPS support (endianness detection)
- Fixed code with exotic architectures (oversized integers)
-
new in sphlib-1.1:
- Fixed bug in Panama implementation (some special padding cases)
- Added RadioGatun[32] and RadioGatun[64] (C)
- Added SHABAL-192/224/256/384/512 (C and Java)
- Added API for fractional bits on some functions (MD5, SHA-0, SHA-1, SHA-224/256/384/512 and SHABAL)
- Added compatibility layer for the NIST SHA-3 competition API