Building Images for Vanilla PluginΒΆ

In this document you will find instruction on how to build Ubuntu, Fedora, and CentOS images with Apache Hadoop versions 1.x.x and 2.x.x.

As of now the vanilla plugin works with images with pre-installed versions of Apache Hadoop. To simplify the task of building such images we use Disk Image Builder.

Disk Image Builder builds disk images using elements. An element is a particular set of code that alters how the image is built, or runs within the chroot to prepare the image.

Elements for building vanilla images are stored in Sahara extra repository


Sahara requires images with cloud-init package installed:

To create vanilla images follow these steps:

  1. Clone repository “” locally.

  2. Run the script.

    You can run the script in any directory (for example, in your home directory). By default this script will attempt to create 6 cloud images, 2 each of Ubuntu, Fedora, and CentOS with versions 1 and 2 of Apache Hadoop. This script must be run with root privileges.

    sudo bash
    This scripts will update your system and install required packages.
    • kpartx
    • qemu
    Then it will clone the repositories “” and “” and export nessesary parameters.
    • DIB_HADOOP_VERSION - version of Hadoop to install
    • JAVA_DOWNLOAD_URL - download link for JDK (tarball or bin)
    • OOZIE_DOWNLOAD_URL - download link for OOZIE (we have built
    Oozie libs here:
    • HIVE_VERSION - version of Hive to install (currently supports only 0.11.0)
    • ubuntu_image_name
    • fedora_image_name
    • DIB_IMAGE_SIZE - parameter that specifies a volume of hard disk of instance. You need to specify it only for Fedora because Fedora doesn’t use all available volume
    • DIB_COMMIT_ID - latest commit id of diksimage-builder project
    • SAHARA_ELEMENTS_COMMIT_ID - latest commit id of sahara-image-elements project

    NOTE: If you don’t want to use default values, you should edit this script and set your values of parameters.

    Then it will create a series of cloud images with hadoop, hive, oozie, mysql, and swift_hadoop elements that install all the necessary packages and configure them. You will find these images in current directory.

For finer control of see the official documentation or run $ -h.

Previous topic

Swift Integration

Next topic

Development Guidelines

This Page