Hazy components are provided as Docker (or OCI) container images in order to simplify their installation and use.
Hazy runs equally well on bare-metal or within a Virtual Machine (VM).
The hardware requirements below are for guidance only. Your installation may need more or fewer resources depending on your requirements and data.
- Processor: x86-64, 1+ cores.
- RAM: >= 4 GiB
- Storage: >= 1 TiB (with backup)
- Processor: x86-64, 4+ cores.
- RAM: >= 64GiB
Hazy containers are self-contained and require no software beyond an OCI Container Runtime.
Hazy currently supports the following Container Runtime implementations:
Docker (https://docker.com) version 19+.
By default Docker runs all containers as
root, but newer Docker versions also support rootless containers
Podman (https://podman.io) version 1.6+.
This is the default runtime for RHEL installations. Podman supports both root-full and root-less containers.
And the following container orchestration systems:
Storage requirements are very installation specific, but a 1TiB partition should be considered a good baseline. The Hub stores:
- Synthesiser container images. (~1GiB per image)
- Generator Models. (size very dependent on data, ~ 1Mib-1GiB per Model)
- Database State and snapshots (~1GiB)
Over time the number of Synthesiser images and Generator Models grows linearly.
The Hub, as a long running service with persistent state, requires its data disk to be regularly backed up to prevent data-loss.
The Synthesiser has no persistent state but does require available storage to read in source data and write out trained Generator Models.
Hazy is designed to work entirely within an on-premise installation or within a private cloud environment and to respect the expected security constraints of production environments. As such, Hazy applications do not require internet access to function.
To function the Hub must be on the same network as the users requiring access to the Generator Models.
The Synthesiser requires access to the source data so must be installed within the same network partition as the production servers or data lake.