JSONIDentification tool and ruleset. jsonid
can be downloaded
from pypi.org.
jsonid
borrows from the Python approach to ask forgiveness rather than
permission (EAFP) to attempt to open every object it scans and see if it
parses as JSON. If it doesn't, we move along. If it does, we then have an
opportunity to identify the characteristics of the JSON we have opened.
Python being high-level also provides an easier path to processing files and parsing JSON quickly with very little other knowledge required of the underlying data structure.
Consider these equivalent forms:
{
"key 1": "value",
"key 2": "value"
}
{
"key 2": "value",
"key 1": "value"
}
PRONOM signatures are not expressive enough for complicated JSON objects.
If I want DROID to find key 1
I have to use a wildcard, so I would write
something like:
BOF: "7B*226B6579203122"
EOF: "7D"
But if I then want to match on key 2
as well as key 1
things start getting
complicated as they aren't guaranteed by the JSON specification to be in the
same "position" (if we think about order visually). When other keys are used in
the object they aren't even guaranteed to be next to one another.
This particular example is a 'map' object whose most important property is consistent retrieval of information through its "keys". Further complexity can be added when we are dealing with maps embedded in a "list" or "array", or simply just maps of arbitrary depth.
jsonid
tries to compensate for JSON's complexities by using the format's
own strengths to parse binary data as JSON and then if is successful,
use a JSON-inspired grammar to describe keys and key-value pairs as "markers"
that can potentially identify the JSON objects that we are looking at.
Certainly narrow down the potential instances of JSON objects that we might
be looking at.
To begin, jsonid
should identify JSON files on your system as JSON.
That's already a pretty good position to be in.
The ruleset should then allow you to identify a decent number of JSON objects, especially those that have a well-defined structure. Examples we have in the registry data include things like ActivityPub streams, RO-CRATE metadata, IIIF API data and so on.
If the ruleset works for JSON we might be able to apply it to other formats that can represent equivalent data structures in the future such as YAML, and TOML.
jsonid
currently defines a small set of rules that help us to identify JSON
documents.
The rules are described in their own data-structures. The structures are processed as a list (they need not necessarily be in order) and each must match for a given set of ruls to determine what kind of JSON document we might be looking at.
jsonid
can identify the existence of information but you can also use
wildcards and provide some negation as required, e.g. to remove
false-positives between similar JSON entities.
rule | meaning |
---|---|
INDEX | index (from which to read when structure is an array) |
GOTO | goto key (read key at given key) |
KEY | key to read |
CONTAINS | value contains string |
STARTSWITH | value startswith string |
ENDSWITH | value endswith string |
IS | value matches exactly |
REGEX | value matches a regex pattern |
EXISTS | key exists |
NOEXIST | key doesn't exists |
ISTYPE | key is a specific type (string, number, dict, array) |
Stored in a list within a RegistryEntry
object, they are then processed
in order.
For example:
[
{ "KEY": "name", "IS": "value" },
{ "KEY": "schema", "CONTAINS": "/schema/version/1.1/" },
{ "KEY": "data", "IS": { "more": "data" } },
]
All rules need to match for a positive ID.
NB.:
jsonid
is a work-in-progress and requires community input to help determine the grammar in its fullness and so there is a lot of opportunity to add/remove to these methods as its development continues. Additionally, help formalizing the grammar/ruleset would be greatly appreciated 🙏.
The ruleset has been developed using test-driven-development practices (TDD) and the current set of tests can be reviewed in the repository's test folder. More tests should be added, in general, and over time.
A temporary "registry" module is used to store JSON markers.
The registry is a work in progress and must be exported and
rewritten somewhere more centralized (and easier to manage) if jsonid
can
prove useful to the communities that might use it (see notes on PRONOM below).
The registry web-page is here:
The registry's source is here:
RegistryEntry(
identifier="id0009",
name=[{"@en": "JSON-LD (generic)"}],
markers=[
{"KEY": "@context", "EXISTS": None},
{"KEY": "id", "EXISTS": None},
],
),
Pseudo code: Test for the existence of keys:
@context
andid
in the primary JSON object.
RegistryEntry(
identifier="id0024",
name=[{"@en": "tika recursive metadata"}],
markers=[
{"INDEX": 0, "KEY": "Content-Length", "EXISTS": None},
{"INDEX": 0, "KEY": "Content-Type", "EXISTS": None},
{"INDEX": 0, "KEY": "X-TIKA:Parsed-By", "EXISTS": None},
{"INDEX": 0, "KEY": "X-TIKA:parse_time_millis", "EXISTS": None},
],
Pseudo code: Test for the existence of keys:
Content-Length
,Content-Type
,X-TIKA:Parsed-By
andX-TIKA:parse_time_millis
in thezeroth
(first) JSON object where the primary document is a list of JSON objects.
RegistryEntry(
identifier="id0012",
name=[{"@en": "sops encrypted secrets file"}],
markers=[
{"KEY": "sops", "EXISTS": None},
{"GOTO": "sops", "KEY": "kms", "EXISTS": None},
{"GOTO": "sops", "KEY": "pgp", "EXISTS": None},
],
),
Pseudo code: Test for the existence of keys
sops
in the primary JSON object.Goto the
sops
key and test for the existence of keys:kms
andpgp
within thesops
object/value.
The plan is to allow local rules to be run alongside the global ruleset. I expect this will be a bit further down the line when the ruleset and metaddata is more stabilised.
Ideally jsonid
can generate evidence enough to warrant the creration of
PRONOM IDs that can then be referenced in the jsonid
output.
Evantually, PRONOM or a PRONOM-like tool might host an authoritative version
of the jsonid
registry.
For ease of development, the utility currently outputs yaml
. The structure
is still very fluid, and will also vary depending on the desired level of
detail in the registry, e.g. there isn't currently a lot of information about
the contents beyond a basic title and identifier.
E.g.:
---
jsonid: 0.0.0
scandate: 2025-04-21T18:40:48Z
---
file: integration_files/plain.json
additional:
- '@en': data is dict type
depth: 1
documentation:
- archive_team: http://fileformats.archiveteam.org/wiki/JSON
identifiers:
- rfc: https://datatracker.ietf.org/doc/html/rfc8259
- pronom: http://www.nationalarchives.gov.uk/PRONOM/fmt/817
- loc: https://www.loc.gov/preservation/digital/formats/fdd/fdd000381.shtml
- wikidata: https://www.wikidata.org/entity/Q2063
mime:
- application/json
name:
- '@en': JavaScript Object Notation (JSON)
---
The structure should become more concrete as jsonid
is formalized.
jsonid
provides an analysis mechanism to help developers of identifiers. It
might also help users talk about interesting properties about the objects
being analysed, and provide consistent fingerprinting for data that has
different byte-alignment but is otherwise identical.
NB.: Comments on existing statistics or ideas for new ones are appreciated.
{
"content_length": 329,
"number_of_lines": 32,
"line_warning": false,
"top_level_keys_count": 4,
"top_level_keys": [
"key1",
"key2",
"key3",
"key4"
],
"top_level_types": [
"list",
"map",
"list",
"list"
],
"depth": 8,
"heterogeneous_list_types": true,
"fingerprint": {
"unf": "UNF:6:sAsKNmjOtnpJtXi3Q6jVrQ==",
"cid": "bafkreibho6naw5r7j23gxu6rzocrud4pc6fjsnteyjveirtnbs3uxemv2u"
},
"encoding": "UTF-8"
}
UTF-16 can be difficult to read as UTF-16 uses two bytes per every one, e.g.
..{.".a.".:. .".b.".}.
is simply {"a": "b"}
. The utility json2json.py
in the utils folder will output UTF-16 as UTF-8 so that signatures can be
more easily derived. A signature derived for UTF-16 looks exactly the same
as UTF-8.
json2json
can be called from the command line when installed via pip, or
find it in src.utils.
Dev docs are available.
Setup a virtual environment venv
and install the local development
requirements as follows:
python3 -m venv venv
source venv/bin/activate
python -m pip install -r requirements/local.txt
python -m tox
python -m tox -e py3
python -m tox -e linting
Pre-commit can be used to provide more feedback before committing code. This reduces reduces the number of commits you might want to make when working on code, it's also an alternative to running tox manually.
To set up pre-commit, providing pip install
has been run above:
pre-commit install
This repository contains a default number of pre-commit hooks, but there may be others suited to different projects. A list of other pre-commit hooks can be found here.
The justfile
contains helper functions for packaging and release.
Run just help
for more information.
Packaging consumes the metadata in pyproject.toml
which helps to describe
the project on the official pypi.org repository. Have a look at the
documentation and comments there to help you create a suitably descriptive
metadata file.
Versioning in Python can be hit and miss. You can label versions for yourself, but to make it reliaable, as well as meaningful is should be controlled by your source control system. We assume git, and versions can be created by tagging your work and pushing the tag to your git repository, e.g. to create a release candidate for version 1.0.0:
git tag -a 1.0.0-rc.1 -m "release candidate for 1.0.0"
git push origin 1.0.0-rc.1
When you build, a package will be created with the correct version:
just package-source
### build process here ###
Successfully built python_repo_jsonid-1.0.0rc1.tar.gz and python_repo_jsonid-1.0.0rc1-py3-none-any.whl
To create a python wheel for testing locally, or distributing to colleagues run:
just package-source
A tar
and whl
file will be stored in a dist/
directory. The whl
file
can be installed as follows:
pip install <your-package>.whl
Publishing for public use can be achieved with:
just package-upload-test
orjust package-upload
just-package-upload-test
will upload the package to test.pypi.org
which provides a way to look at package metadata and documentation and ensure
that it is correct before uploading to the official pypi.org
repository using just package-upload
.