Packer
Artifice Post-Processor
Official
Type: artifice
Artifact BuilderId: packer.post-processor.artifice
The artifice post-processor overrides the artifact list from an upstream builder or post-processor. All downstream post-processors will see the new artifacts you specify.
After overriding the artifact with artifice, you can use it with other post-processors, including most of the core post-processors and third-party post-processors.
A major benefit of this is that you can modify builder artifacts using shell-local and pass those modified artifacts into post-processors that may not have worked with the original builder. For example, maybe you want to export a Docker container from an amazon-ebs builder and then use Docker-push to put that Docker container into your Docker Hub account.
Artifice allows you to use the familiar packer workflow to create a fresh, stateless build environment for each build on the infrastructure of your choosing. You can use this to build just about anything: buildpacks, containers, jars, binaries, tarballs, msi installers, and more.
Please note that the artifice post-processor will not delete your old artifact files, even if it removes them from the artifact. If you want to delete the old artifact files, you can use the shell-local post-processor to do so.
Workflow
Artifice helps you tie together a few other packer features:
- A builder, which spins up a VM (or container) to build your artifact
- A provisioner, which performs the steps to create your artifact
- A file provisioner, which downloads the artifact from the VM
- The artifice post-processor, which identifies which files have been downloaded from the VM
- Additional post-processors, which push the artifact to Docker hub, etc.
You will want to perform as much work as possible inside the VM. Ideally the only other post-processor you need after artifice is one that uploads your artifact to the appropriate repository.
Configuration
The configuration allows you to specify which files comprise your artifact.
Required:
files
(array of strings) - A list of files that comprise your artifact. These files must exist on your local disk after the provisioning phase of packer is complete. These will replace any of the builder's original artifacts (such as a VM snapshot).
Optional:
keep_input_artifact
(boolean) - if true, do not delete the original artifact files after creating your new artifact. Defaults to true.
Example Configuration
This minimal example:
- Spins up a cloned VMware virtual machine
- Installs a consul release
- Downloads the consul binary
- Packages it into a
.tar.gz
file - Uploads it to S3.
VMX is a fast way to build and test locally, but you can easily substitute another builder.
{
"builders": [
{
"type": "vmware-vmx",
"source_path": "/opt/ubuntu-1404-vmware.vmx",
"ssh_username": "vagrant",
"ssh_password": "vagrant",
"shutdown_command": "sudo shutdown -h now",
"headless": "true",
"skip_compaction": "true"
}
],
"provisioners": [
{
"type": "shell",
"inline": [
"sudo apt-get install -y python-pip",
"sudo pip install ifs",
"sudo ifs install consul --version=0.5.2"
]
},
{
"type": "file",
"source": "/usr/local/bin/consul",
"destination": "consul",
"direction": "download"
}
],
"post-processors": [
[
{
"type": "artifice",
"files": ["consul"]
},
{
"type": "compress",
"output": "consul-0.5.2.tar.gz"
},
{
"type": "shell-local",
"inline": [
"/usr/local/bin/aws s3 cp consul-0.5.2.tar.gz s3://<s3 path>"
]
}
]
]
}
Notice that there are two sets of square brackets in the post-processor section. This creates a post-processor chain, where the output of the proceeding artifact is passed to subsequent post-processors. If you use only one set of square braces the post-processors will run individually against the build artifact (the vmx file in this case) and it will not have the desired result.
{
"post-processors": [
[ // <--- Start post-processor chain
{
"type": "artifice",
"files": ["consul"]
},
{
"type": "compress",
...
}
], // <--- End post-processor chain
{
"type":"compress" // <-- Standalone post-processor
}
]
}
You can create multiple post-processor chains to handle multiple builders (for example, building Linux and Windows binaries during the same build).