diff options
Diffstat (limited to 'unmaintained/clusters')
-rw-r--r-- | unmaintained/clusters/cephclient.morph | 20 | ||||
-rw-r--r-- | unmaintained/clusters/example-ceph-cluster-on-openstack.morph | 55 | ||||
-rw-r--r-- | unmaintained/clusters/example-swift-storage-cluster.morph | 62 | ||||
-rw-r--r-- | unmaintained/clusters/jetson-upgrade.morph | 17 | ||||
-rw-r--r-- | unmaintained/clusters/moonshot-m2-armv8b64.morph | 56 | ||||
-rw-r--r-- | unmaintained/clusters/moonshot-pxe-armv8b64.morph | 30 | ||||
-rw-r--r-- | unmaintained/clusters/moonshot-pxe-armv8l64.morph | 22 | ||||
-rw-r--r-- | unmaintained/clusters/openstack-one-node-swift.morph | 142 | ||||
-rw-r--r-- | unmaintained/clusters/openstack-one-node.morph | 104 | ||||
-rw-r--r-- | unmaintained/clusters/openstack-three-node-installer.morph | 238 | ||||
-rw-r--r-- | unmaintained/clusters/openstack-two-node-installer.morph | 199 | ||||
-rw-r--r-- | unmaintained/clusters/sdk-example-cluster.morph | 46 | ||||
-rw-r--r-- | unmaintained/clusters/upgrade-devel.morph | 39 | ||||
-rw-r--r-- | unmaintained/clusters/xfce-system-x86_64-deploy.morph | 23 | ||||
-rw-r--r-- | unmaintained/clusters/zookeeper.morph | 21 |
15 files changed, 0 insertions, 1074 deletions
diff --git a/unmaintained/clusters/cephclient.morph b/unmaintained/clusters/cephclient.morph deleted file mode 100644 index ad680a76..00000000 --- a/unmaintained/clusters/cephclient.morph +++ /dev/null @@ -1,20 +0,0 @@ -name: cephclient -kind: cluster -systems: -- morph: unmaintained/systems/ceph-service-x86_64-generic.morph - deploy: - ceph-node-virtualbox-image: - type: extensions/virtualbox-ssh - SYSTEM: systems/ceph-service-x86_64-generic.morph - location: vbox+ssh://user@machine/ChefNode4/home/user/chefnode4.vdi - # HOST_IPADDR and NETMASK should be set to the IP address and netmask of the virtualbox host on the host-only interface. - #HOST_IPADDR: 10.0.100.100 - #NETMASK: 255.255.255.0 - - # This is an example of how to configure the three interfaces necessary to support ceph in the BCPC configuration. - #NETWORK_CONFIG: lo:loopback;enp0s3:static,address=10.0.100.14,netmask=255.255.255.0;enp0s8:static,address=172.16.100.14,netmask=255.255.255.0;enp0s9:static,address=192.168.100.14,netmask=255.255.255.0 - DISK_SIZE: 8G - HOSTNAME: CephNode4 - - # You must install authorized_keys in chef/root/.ssh/ before this will work. - INSTALL_FILES: install-files/chef/manifest diff --git a/unmaintained/clusters/example-ceph-cluster-on-openstack.morph b/unmaintained/clusters/example-ceph-cluster-on-openstack.morph deleted file mode 100644 index 73f6729c..00000000 --- a/unmaintained/clusters/example-ceph-cluster-on-openstack.morph +++ /dev/null @@ -1,55 +0,0 @@ -name: example-ceph-cluster-on-openstack -kind: cluster -description: | - This cluster morphology will deploy a 3 node ceph storage cluster to an - openstack server. - - It was written for use with the following guide on the baserock wiki: - http://wiki.baserock.org/guides/ceph-cluster-deploy/ - - See this guide for more information. - - See the ceph.configure file for more informatiion on the ceph - specific fields used in this cluster morphology. - -systems: -- morph: unmaintained/systems/ceph-service-x86_64-generic.morph - deploy: - ceph-mon-0-openstack: - # openstack info - <<: &common-config - type: openstack - location: http://<BASEROCK_OPENSTACK_SERVER_IP>:5000/v2.0/ - OPENSTACK_USER: demo - OPENSTACK_PASSWORD: demo - OPENSTACK_TENANT: demo - CLOUD_INIT: true - KERNEL_ARGS: console=ttyS0 console=tty0 - CEPH_CLUSTER: ceph - CEPH_CONF: ceph.conf - CEPH_CLIENT_ADMIN: ceph.client.admin.keyring - CEPH_MON_IP: <CEPH_MON_IP> - CEPH_CLUSTER_FSID: <CEPH_CLUSTER_FSID> - - OPENSTACK_IMAGENAME: mon-0-nd - DISK_SIZE: 10G - # Ceph info - HOSTNAME: mon-0 - CEPH_MON: - ceph-osd-0-openstack: - <<: *common-config - OPENSTACK_IMAGENAME: osd-0 - DISK_SIZE: 7G - # ceph info - HOSTNAME: osd-0 - CEPH_OSD: - CEPH_OSD_STORAGE_DEV: /dev/vdb - ceph-osd-1-openstack: - <<: *common-config - OPENSTACK_IMAGENAME: osd-1 - DISK_SIZE: 7G - # ceph info - HOSTNAME: osd-1 - - CEPH_OSD: - CEPH_OSD_STORAGE_DEV: /dev/vdb diff --git a/unmaintained/clusters/example-swift-storage-cluster.morph b/unmaintained/clusters/example-swift-storage-cluster.morph deleted file mode 100644 index 42f52d8a..00000000 --- a/unmaintained/clusters/example-swift-storage-cluster.morph +++ /dev/null @@ -1,62 +0,0 @@ -name: example-swift-storage-cluster -kind: cluster -systems: -- morph: unmaintained/systems/swift-system-x86_64.morph - deploy-defaults: - INSTALL_FILES: install-files/swift/manifest - - CONTROLLER_HOST_ADDRESS: <controller host address> - - SWIFT_PART_POWER: 10 - SWIFT_REPLICAS: 3 - SWIFT_MIN_PART_HOURS: 1 - - SWIFT_STORAGE_DEVICES: - - {ip: <node0 management ip>, device: sdb, weight: 100} - - {ip: <node0 management ip>, device: sdc, weight: 100} - - {ip: <node0 management ip>, device: sdd, weight: 100} - - {ip: <node1 management ip>, device: sdb, weight: 100} - - {ip: <node1 management ip>, device: sdc, weight: 100} - - {ip: <node1 management ip>, device: sdd, weight: 100} - - # This value can be any random string or number - # but each node in your swift cluster must have the same value - SWIFT_REBALANCE_SEED: 3828 - - # NOTE: Replace SWIFT_HASH_PATH_PREFIX and SWIFT_HASH_PATH_SUFFIX - # with your own unique values, - # - # `openssl rand -hex 10' can be used to generate unique values - # - # These values should be kept secret, do not lose them. - # - SWIFT_HASH_PATH_PREFIX: 041fc210e4e1d333ce1d - SWIFT_HASH_PATH_SUFFIX: 4d6f5362a356dda7fb7d - - FSTAB_SDB: /dev/sdb /srv/node/sdb xfs defaults,user,rw 0 0 - FSTAB_SDC: /dev/sdc /srv/node/sdc xfs defaults,user,rw 0 0 - FSTAB_SDD: /dev/sdd /srv/node/sdd xfs defaults,user,rw 0 0 - - deploy: - node0: - type: extensions/kvm - location: kvm+ssh://user@host/swift-storage-0/home/user/swift-storage-0.img - DISK_SIZE: 10G - RAM_SIZE: 1G - VCPUS: 1 - HOSTNAME: swift-storage-0 - NIC_CONFIG: network=default - NETWORK_CONFIG: ens3:static,address=<node0 management ip>,netmask=255.255.255.0 - MANAGEMENT_INTERFACE_IP_ADDRESS: <node0 management ip> - ATTACH_DISKS: /dev/node0_sdb:/dev/node0_sdc:/dev/node0_sdd - node1: - type: extensions/kvm - location: kvm+ssh://user@host/swift-storage-1/home/user/swift-storage-1.img - DISK_SIZE: 10G - RAM_SIZE: 1G - VCPUS: 1 - HOSTNAME: swift-storage-1 - NIC_CONFIG: network=default - NETWORK_CONFIG: ens3:static,address=<node1 management ip>,netmask=255.255.255.0 - MANAGEMENT_INTERFACE_IP_ADDRESS: <node1 management ip> - ATTACH_DISKS: /dev/node1_sdb:/dev/node1_sdc:/dev/node1_sdd diff --git a/unmaintained/clusters/jetson-upgrade.morph b/unmaintained/clusters/jetson-upgrade.morph deleted file mode 100644 index fb1fc1b4..00000000 --- a/unmaintained/clusters/jetson-upgrade.morph +++ /dev/null @@ -1,17 +0,0 @@ -name: jetson-upgrade -kind: cluster -systems: -- morph: unmaintained/systems/devel-system-armv7lhf-jetson.morph - deploy-defaults: - BOOT_DEVICE: /dev/mmcblk0p1 - ROOT_DEVICE: /dev/mmcblk0p2 - DTB_PATH: boot/tegra124-jetson-tk1.dtb - BOOTLOADER_CONFIG_FORMAT: extlinux - BOOTLOADER_INSTALL: none - KERNEL_ARGS: cma=256M console=tty0 console=ttyS0,115200n8 no_console_suspend=1 - nouveau.pstate=1 - FSTAB_SRC: LABEL=src /src auto defaults,rw,noatime,nofail 0 2 - deploy: - self: - upgrade-type: extensions/ssh-rsync - upgrade-location: root@127.0.0.1 diff --git a/unmaintained/clusters/moonshot-m2-armv8b64.morph b/unmaintained/clusters/moonshot-m2-armv8b64.morph deleted file mode 100644 index 9c68248b..00000000 --- a/unmaintained/clusters/moonshot-m2-armv8b64.morph +++ /dev/null @@ -1,56 +0,0 @@ -name: moonshot-m2-armv8b64 -kind: cluster -description: | - Install a build armv8b64 system into the M.2 SSD storage of an HP - Moonshot node, using a combination of the pxeboot.write extension and - the installer system. -systems: -- morph: systems/installer-system-armv8b64.morph - deploy: - installer: - type: extensions/pxeboot - location: 14:58:d0:57:7f:42 - PXEBOOT_MODE: existing-server - PXEBOOT_CONFIG_TFTP_ADDRESS: sftp://192.168.0.1/srv/nfsboot/tftp/ - PXEBOOT_ROOTFS_RSYNC_ADDRESS: rsync://192.168.0.1/srv/nfsboot/ - PXEBOOT_PXE_REBOOT_COMMAND: | - ssh Administrator@10.0.1.10 set node power off force c31n1 - ssh Administrator@10.0.1.10 set node boot pxe c31n1 - # Nodes are powered on twice as sometimes powering them on - # once is not enough - ssh Administrator@10.0.1.10 set node power on c31n1 - ssh Administrator@10.0.1.10 set node power on c31n1 - PXEBOOT_REBOOT_COMMAND: | - ssh Administrator@10.0.1.10 set node power off force c31n1 - ssh Administrator@10.0.1.10 set node boot m.2 c31n1 - ssh Administrator@10.0.1.10 set node power on c31n1 - ssh Administrator@10.0.1.10 set node power on c31n1 - - INSTALLER_TARGET_STORAGE_DEVICE: /dev/sda - INSTALLER_ROOTFS_TO_INSTALL: /rootfs - INSTALLER_POST_INSTALL_COMMAND: | - while : ; do - echo "enter 'installed' in your deployment machine to finish the installation" - sleep 2 - done - INSTALLER_CREATE_BOOT_PARTITION: true - - HOSTNAME: installer-system-c31n1 - DTB_PATH: boot/m400-1003.dtb - KERNEL_ARGS: console=ttyS0,9600n8r init=/usr/lib/baserock-installer/installer - INSTALL_FILES: install-files/moonshot/manifest - MOONSHOT_KERNEL: true - subsystems: - - morph: unmaintained/systems/devel-system-armv8b64.morph - deploy: - to-install: - type: extensions/sysroot - location: /rootfs - HOSTNAME: baserock-c31n1 - DTB_PATH: boot/m400-1003.dtb - INSTALL_FILES: install-files/moonshot/manifest - MOONSHOT_KERNEL: true - BOOT_DEVICE: /dev/sda1 - ROOT_DEVICE: /dev/sda2 - BOOTLOADER_CONFIG_FORMAT: extlinux - BOOTLOADER_INSTALL: none diff --git a/unmaintained/clusters/moonshot-pxe-armv8b64.morph b/unmaintained/clusters/moonshot-pxe-armv8b64.morph deleted file mode 100644 index d97dcbef..00000000 --- a/unmaintained/clusters/moonshot-pxe-armv8b64.morph +++ /dev/null @@ -1,30 +0,0 @@ -name: moonshot-pxe-armv8b64 -kind: cluster -description: | - Deploy a big-endian armv8b64 devel system onto an HP Moonshot node - - The system will be configured to boot through PXE from existing DHCP, - TFTP and NFS servers. -systems: -- morph: unmaintained/systems/devel-system-armv8b64.morph - deploy: - netboot: - type: extensions/pxeboot - location: 14:58:d0:57:7f:42 - PXEBOOT_MODE: existing-server - PXEBOOT_CONFIG_TFTP_ADDRESS: sftp://192.168.0.1/srv/nfsboot/tftp/ - PXEBOOT_ROOTFS_RSYNC_ADDRESS: rsync://192.168.0.1/srv/nfsboot/ - PXEBOOT_PXE_REBOOT_COMMAND: | - ssh Administrator@10.0.1.10 set node power off force c31n1 - ssh Administrator@10.0.1.10 set node boot pxe c31n1 - # Nodes are powered on twice as sometimes powering them on - # once is not enough - ssh Administrator@10.0.1.10 set node power on c31n1 - ssh Administrator@10.0.1.10 set node power on c31n1 - PXE_INSTALLER: false - - HOSTNAME: baserock-c31n1 - DTB_PATH: boot/m400-1003.dtb - KERNEL_ARGS: console=ttyS0,9600n8r rw - INSTALL_FILES: install-files/moonshot/manifest - MOONSHOT_KERNEL: true diff --git a/unmaintained/clusters/moonshot-pxe-armv8l64.morph b/unmaintained/clusters/moonshot-pxe-armv8l64.morph deleted file mode 100644 index 531c7428..00000000 --- a/unmaintained/clusters/moonshot-pxe-armv8l64.morph +++ /dev/null @@ -1,22 +0,0 @@ -name: moonshot-pxe-armv8l64 -kind: cluster -description: | - Deploy an armv8l64 devel system into a HP Moonshot node - - The system will be configured to boot through PXE from existing DHCP, - TFTP and NFS servers. -systems: -- morph: unmaintained/systems/devel-system-armv8l64.morph - deploy: - netboot: - type: extensions/pxeboot - location: 14:58:d0:57:7f:42 - PXEBOOT_MODE: existing-server - PXEBOOT_CONFIG_TFTP_ADDRESS: sftp://192.168.0.1/srv/nfsboot/tftp/ - PXEBOOT_ROOTFS_RSYNC_ADDRESS: rsync://192.168.0.1/srv/nfsboot/ - KERNEL_ARGS: console=ttyS0,9600n8r rw - DTB_PATH: boot/m400-1003.dtb - HOSTNAME: baserock-m400-node31 - MOONSHOT_KERNEL: true - INSTALL_FILES: install-files/moonshot/manifest - PXE_INSTALLER: false diff --git a/unmaintained/clusters/openstack-one-node-swift.morph b/unmaintained/clusters/openstack-one-node-swift.morph deleted file mode 100644 index b9d6502b..00000000 --- a/unmaintained/clusters/openstack-one-node-swift.morph +++ /dev/null @@ -1,142 +0,0 @@ -name: openstack-one-node-swift -kind: cluster -description: | - This is a cluster morphology for deploying a x86_64 OpenStack system - all-in-one-node. - - Requirements to be able to run and test the system: - - - DISK_SIZE should be bigger than 5G - - The system has to have available at least 4G of RAM, but once - you start instantiating VMs you will need more. - - The IP of the system can't change, and you need to know it beforehand, - that is, the system needs a static IP address. - - This cluster is configurable, but with the following constraints: - - - The hostname in RABBITMQ_HOST has to match CONTROLLER_HOST_ADDRESS, - and HOST_CONTROLLER. - - HOSTS_CONTROLLER is only needed if the hostname (see previous point) - is not a FQDN. - - The IP listed in MANAGEMENT_INTERFACE_IP_ADDRESS has to match the one - used in HOSTS_CONTROLLER. - - CINDER_DEVICE should be a path to a storage device ready to be - used/formated for cinder data. - - EXTERNAL_INTERFACE is required when the system has more than one network - interface. - - You can also have a look at the following suggestions: - - - NOVA_VIRT_TYPE can be either 'kvm' or 'qemu', depending on where the - system is being deployed to. - - We recommend changing all the PASSWORDs variables, also the - KEYSTONE_TEMPORARY_ADMIN_TOKEN, METADATA_PROXY_SHARED_SECRET and - METERING_SECRET - - Setting NOVA_BAREMETAL_SCHEDULING with an YAML truth value will configure - Nova to schedule baremetal machines through the Ironic driver, instead of - sheduling virtual machines. - -systems: -- morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - release: - type: extensions/rawdisk - location: baserock-openstack-system-x86_64.img - DISK_SIZE: 10G - INSTALL_FILES: install-files/openstack/manifest install-files/swift/manifest - - HOSTNAME: onenode - - ######################################################################### - ## Swift config options - ######################################################################### - - SWIFT_CONTROLLER: true - - SWIFT_ADMIN_PASSWORD: insecure - - SWIFT_PART_POWER: 10 - SWIFT_REPLICAS: 3 - SWIFT_MIN_PART_HOURS: 1 - - SWIFT_STORAGE_DEVICES: - - {ip: <storage node 0 management ip>, device: sdb, weight: 100} - - {ip: <storage node 0 management ip>, device: sdc, weight: 100} - - {ip: <storage node 0 management ip>, device: sdd, weight: 100} - - {ip: <storage node 1 management ip>, device: sdb, weight: 100} - - {ip: <storage node 1 management ip>, device: sdc, weight: 100} - - {ip: <storage node 1 management ip>, device: sdd, weight: 100} - - # This value can be any random string or number - # but each node in your swift cluster must have the same values - SWIFT_REBALANCE_SEED: 3828 - - # NOTE: Replace SWIFT_HASH_PATH_PREFIX and SWIFT_HASH_PATH_SUFFIX - # with your own unique values, - # - # `openssl rand -hex 10' can be used to generate unique values - # - # These values should be kept secret, do not lose them. - # - SWIFT_HASH_PATH_PREFIX: 041fc210e4e1d333ce1d - SWIFT_HASH_PATH_SUFFIX: 4d6f5362a356dda7fb7d - - ######################################################################### - - RABBITMQ_HOST: onenode - RABBITMQ_PORT: 5672 - RABBITMQ_USER: rabbitmq - RABBITMQ_PASSWORD: veryinsecure - - CONTROLLER_HOST_ADDRESS: onenode - MANAGEMENT_INTERFACE_IP_ADDRESS: <management ip> - - KEYSTONE_ENABLE_SERVICE: true - KEYSTONE_TEMPORARY_ADMIN_TOKEN: 22f3aa1cf538e3f6d5e8 - KEYSTONE_ADMIN_PASSWORD: veryinsecure - KEYSTONE_DB_USER: keystoneDB - KEYSTONE_DB_PASSWORD: veryinsecure - - GLANCE_ENABLE_SERVICE: true - GLANCE_SERVICE_USER: glance - GLANCE_SERVICE_PASSWORD: veryinsecure - GLANCE_DB_USER: glanceDB - GLANCE_DB_PASSWORD: veryinsecure - - NOVA_SERVICE_USER: nova - NOVA_SERVICE_PASSWORD: veryinsecure - NOVA_DB_USER: novaDB - NOVA_DB_PASSWORD: veryinsecure - NOVA_VIRT_TYPE: qemu - NOVA_BAREMETAL_SCHEDULING: false - - CINDER_SERVICE_USER: cinder - CINDER_SERVICE_PASSWORD: veryinsecure - CINDER_DB_USER: cinderDB - CINDER_DB_PASSWORD: veryinsecure - # Storage device to be used by Cinder - CINDER_DEVICE: /dev/sdb - - NEUTRON_SERVICE_USER: neutron - NEUTRON_SERVICE_PASSWORD: veryinsecure - NEUTRON_DB_USER: neutronDB - NEUTRON_DB_PASSWORD: veryinsecure - METADATA_PROXY_SHARED_SECRET: novaneutronmetasecret - - IRONIC_ENABLE_SERVICE: true - IRONIC_SERVICE_USER: ironic - IRONIC_SERVICE_PASSWORD: veryinsecure - IRONIC_DB_USER: ironicDB - IRONIC_DB_PASSWORD: veryinsecure - - CEILOMETER_SERVICE_USER: ceilometer - CEILOMETER_SERVICE_PASSWORD: veryinsecure - CEILOMETER_DB_USER: ceilometerDB - CEILOMETER_DB_PASSWORD: veryinsecure - METERING_SECRET: insecureceilometersecret - - HOSTS_CONTROLLER: <management ip> onenode - - # Network interface to be used, only needed if there are more - # than one available. - # EXTERNAL_INTERFACE: eno1 diff --git a/unmaintained/clusters/openstack-one-node.morph b/unmaintained/clusters/openstack-one-node.morph deleted file mode 100644 index ddecb6fc..00000000 --- a/unmaintained/clusters/openstack-one-node.morph +++ /dev/null @@ -1,104 +0,0 @@ -name: openstack-one-node -kind: cluster -description: | - This is a cluster morphology for deploying a x86_64 OpenStack system - all-in-one-node. - - Requirements to be able to run and test the system: - - - DISK_SIZE should be bigger than 5G - - The system has to have available at least 4G of RAM, but once - you start instantiating VMs you will need more. - - The IP of the system can't change, and you need to know it beforehand, - that is, the system needs a static IP address. - - This cluster is configurable, but with the following constraints: - - - The hostname in RABBITMQ_HOST has to match CONTROLLER_HOST_ADDRESS, - and HOST_CONTROLLER. - - HOSTS_CONTROLLER is only needed if the hostname (see previous point) - is not a FQDN. - - The IP listed in MANAGEMENT_INTERFACE_IP_ADDRESS has to match the one - used in HOSTS_CONTROLLER. - - CINDER_DEVICE should be a path to a storage device ready to be - used/formated for cinder data. - - EXTERNAL_INTERFACE is required when the system has more than one network - interface. - - You can also have a look at the following suggestions: - - - NOVA_VIRT_TYPE can be either 'kvm' or 'qemu', depending on where the - system is being deployed to. - - We recommend changing all the PASSWORDs variables, also the - KEYSTONE_TEMPORARY_ADMIN_TOKEN, METADATA_PROXY_SHARED_SECRET and - METERING_SECRET - - Setting NOVA_BAREMETAL_SCHEDULING with an YAML truth value will configure - Nova to schedule baremetal machines through the Ironic driver, instead of - sheduling virtual machines. - -systems: -- morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - release: - type: extensions/rawdisk - location: baserock-openstack-system-x86_64.img - DISK_SIZE: 10G - INSTALL_FILES: install-files/openstack/manifest - - HOSTNAME: onenode - - RABBITMQ_HOST: onenode - RABBITMQ_PORT: 5672 - RABBITMQ_USER: rabbitmq - RABBITMQ_PASSWORD: veryinsecure - - CONTROLLER_HOST_ADDRESS: onenode - MANAGEMENT_INTERFACE_IP_ADDRESS: <management ip> - - KEYSTONE_ENABLE_SERVICE: true - KEYSTONE_TEMPORARY_ADMIN_TOKEN: 22f3aa1cf538e3f6d5e8 - KEYSTONE_ADMIN_PASSWORD: veryinsecure - KEYSTONE_DB_USER: keystoneDB - KEYSTONE_DB_PASSWORD: veryinsecure - - GLANCE_ENABLE_SERVICE: true - GLANCE_SERVICE_USER: glance - GLANCE_SERVICE_PASSWORD: veryinsecure - GLANCE_DB_USER: glanceDB - GLANCE_DB_PASSWORD: veryinsecure - - NOVA_SERVICE_USER: nova - NOVA_SERVICE_PASSWORD: veryinsecure - NOVA_DB_USER: novaDB - NOVA_DB_PASSWORD: veryinsecure - NOVA_VIRT_TYPE: qemu - NOVA_BAREMETAL_SCHEDULING: false - - CINDER_SERVICE_USER: cinder - CINDER_SERVICE_PASSWORD: veryinsecure - CINDER_DB_USER: cinderDB - CINDER_DB_PASSWORD: veryinsecure - # Storage device to be used by Cinder - CINDER_DEVICE: /dev/sdb - - NEUTRON_SERVICE_USER: neutron - NEUTRON_SERVICE_PASSWORD: veryinsecure - NEUTRON_DB_USER: neutronDB - NEUTRON_DB_PASSWORD: veryinsecure - METADATA_PROXY_SHARED_SECRET: novaneutronmetasecret - - IRONIC_ENABLE_SERVICE: true - IRONIC_SERVICE_USER: ironic - IRONIC_SERVICE_PASSWORD: veryinsecure - IRONIC_DB_USER: ironicDB - IRONIC_DB_PASSWORD: veryinsecure - - CEILOMETER_SERVICE_USER: ceilometer - CEILOMETER_SERVICE_PASSWORD: veryinsecure - CEILOMETER_DB_USER: ceilometerDB - CEILOMETER_DB_PASSWORD: veryinsecure - METERING_SECRET: insecureceilometersecret - - # Network interface to be used, only needed if there are more - # than one available. - # EXTERNAL_INTERFACE: eno1 diff --git a/unmaintained/clusters/openstack-three-node-installer.morph b/unmaintained/clusters/openstack-three-node-installer.morph deleted file mode 100644 index 9274f9bf..00000000 --- a/unmaintained/clusters/openstack-three-node-installer.morph +++ /dev/null @@ -1,238 +0,0 @@ -name: openstack-three-node-installer -kind: cluster -description: | - This is a cluster morphology for deploying an installer for an x86_64 - OpenStack system spread across three nodes. - - This cluster creates disk images that may be `dd`'d onto install media to - produce an OpenStack cluster when instantiated. - - Alternatively it may be used to install directly onto a physical disk by - running: - - morph deploy clusters/openstack-three-node-installer.morph \ - network-installer network-installer.location=/dev/vdb - - Substituting network-installer for either compute-installer or - controller-installer will produce different configurations, and it is possible - to substitue /dev/vdb for a different path to a disk image to install to a - different disk image. - - Substitute the values of HOSTNAME, NETWORK_CONFIG, EXTERNAL_INTERFACE, - MANAGEMENT_IP_ADDRESS, CONTROLLER_HOST_ADDRESS, RABBITMQ_HOST and HOSTS_* to - match your hardware and networking configuration. - - Requirements to be able to run and test the system: - - - DISK_SIZE should be bigger than 5G - - The system has to have available at least 4G of RAM, but once - you start instantiating VMs you will need more. - - The IP of the system can't change, and you need to know it beforehand, - that is, the system needs a static IP address. - - This cluster is configurable, but with the following constraints: - - - The hostname in RABBITMQ_HOST has to match CONTROLLER_HOST_ADDRESS, - and HOST_CONTROLLER. - - HOSTS_CONTROLLER is only needed if the hostname (see previous point) - is not a FQDN. - - The IP listed in MANAGEMENT_INTERFACE_IP_ADDRESS has to match the one - used in HOSTS_CONTROLLER. - - CINDER_DEVICE should be a path to a storage device ready to be - used/formated for cinder data. - - EXTERNAL_INTERFACE is required when the system has more than one network - interface. - - You can also have a look at the following suggestions: - - - NOVA_VIRT_TYPE can be either 'kvm' or 'qemu', depending on where the - system is being deployed to. - - We recommend changing all the PASSWORDs variables, also the - KEYSTONE_TEMPORARY_ADMIN_TOKEN, METADATA_PROXY_SHARED_SECRET and - METERING_SECRET. - - Setting NOVA_BAREMETAL_SCHEDULING with an YAML truth value will configure - Nova to schedule baremetal machines through the Ironic driver, instead of - sheduling virtual machines. - -systems: -- morph: systems/installer-system-x86_64.morph - deploy: - network-installer: &installer - type: extensions/rawdisk - location: installer-openstack-network-x86_64.img - KERNEL_ARGS: init=/usr/lib/baserock-installer/installer - DISK_SIZE: 6G - HOSTNAME: installer-x86_64 - INSTALLER_TARGET_STORAGE_DEVICE: /dev/sda - INSTALLER_ROOTFS_TO_INSTALL: /rootfs - INSTALLER_POST_INSTALL_COMMAND: sync; poweroff -f - INITRAMFS_PATH: boot/initramfs.gz - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - network-initramfs: &initramfs - type: extensions/initramfs - location: boot/initramfs.gz - - morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - network-to-install: &stack-node - type: extensions/sysroot - location: rootfs - INSTALL_FILES: install-files/openstack/manifest - INITRAMFS_PATH: boot/initramfs.gz - - HOSTNAME: threenode-network - - RABBITMQ_HOST: threenode-controller.os-mgmt - RABBITMQ_PORT: 5672 - RABBITMQ_USER: rabbitmq - RABBITMQ_PASSWORD: veryinsecure - - # This token needs to be unique and secret - KEYSTONE_ENABLE_SERVICE: false - KEYSTONE_TEMPORARY_ADMIN_TOKEN: 22f3aa1cf538e3f6d5e8 - KEYSTONE_ADMIN_PASSWORD: veryinsecure - KEYSTONE_DB_USER: keystoneDB - KEYSTONE_DB_PASSWORD: veryinsecure - - GLANCE_ENABLE_SERVICE: false - GLANCE_SERVICE_USER: glance - GLANCE_SERVICE_PASSWORD: veryinsecure - GLANCE_DB_USER: glanceDB - GLANCE_DB_PASSWORD: veryinsecure - - NOVA_ENABLE_CONTROLLER: false - NOVA_ENABLE_COMPUTE: false - NOVA_SERVICE_USER: nova - NOVA_SERVICE_PASSWORD: veryinsecure - NOVA_DB_USER: novaDB - NOVA_DB_PASSWORD: veryinsecure - NOVA_VIRT_TYPE: kvm - NOVA_BAREMETAL_SCHEDULING: false - - CINDER_ENABLE_CONTROLLER: false - CINDER_ENABLE_COMPUTE: false - CINDER_ENABLE_STORAGE: false - CINDER_SERVICE_USER: cinder - CINDER_SERVICE_PASSWORD: veryinsecure - CINDER_DB_USER: cinderDB - CINDER_DB_PASSWORD: veryinsecure - # Storage device to be used by Cinder - CINDER_DEVICE: /dev/sdb - - NEUTRON_ENABLE_AGENT: false - NEUTRON_ENABLE_MANAGER: true - NEUTRON_ENABLE_CONTROLLER: false - NEUTRON_SERVICE_USER: neutron - NEUTRON_SERVICE_PASSWORD: veryinsecure - NEUTRON_DB_USER: neutronDB - NEUTRON_DB_PASSWORD: veryinsecure - METADATA_PROXY_SHARED_SECRET: novaneutronmetasecret - - IRONIC_ENABLE_SERVICE: false - IRONIC_SERVICE_USER: ironic - IRONIC_SERVICE_PASSWORD: veryinsecure - IRONIC_DB_USER: ironicDB - IRONIC_DB_PASSWORD: veryinsecure - - CEILOMETER_SERVICE_USER: ceilometer - CEILOMETER_SERVICE_PASSWORD: veryinsecure - CEILOMETER_DB_USER: ceilometerDB - CEILOMETER_DB_PASSWORD: veryinsecure - CEILOMETER_ENABLE_CONTROLLER: false - CEILOMETER_ENABLE_COMPUTE: false - METERING_SECRET: insecureceilometersecret - - CONTROLLER_HOST_ADDRESS: threenode-controller.os-mgmt - MANAGEMENT_INTERFACE_IP_ADDRESS: 10.0.0.1 - - HOSTS_NETWORK: 10.0.0.1 threenode-network.os-mgmt - HOSTS_CONTROL: 10.0.0.2 threenode-controller.os-mgmt - HOSTS_COMPUTE: 10.0.0.3 threenode-compute.os-mgmt - - EXTERNAL_INTERFACE: enp3s0 - NETWORK_CONFIG: enp3s0:dhcp;enp2s0:static,address=10.0.0.1,netmask=255.255.255.0 - SYNC_TIME_WITH_CONTROLLER: true - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - network-to-install-initramfs: *initramfs -- morph: systems/installer-system-x86_64.morph - deploy: - controller-installer: - <<: *installer - location: installer-openstack-controller-x86_64.img - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - controller-initramfs: *initramfs - - morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - controller-to-install: - <<: *stack-node - HOSTNAME: threenode-controller - - KEYSTONE_ENABLE_SERVICE: true - - GLANCE_ENABLE_SERVICE: true - - NOVA_ENABLE_CONTROLLER: true - - CINDER_ENABLE_CONTROLLER: true - CINDER_ENABLE_COMPUTE: false - CINDER_ENABLE_STORAGE: false - - NEUTRON_ENABLE_AGENT: false - NEUTRON_ENABLE_MANAGER: false - NEUTRON_ENABLE_CONTROLLER: true - METADATA_PROXY_SHARED_SECRET: novaneutronmetasecret - - IRONIC_ENABLE_SERVICE: true - - CEILOMETER_ENABLE_CONTROLLER: true - CEILOMETER_ENABLE_COMPUTE: false - - MANAGEMENT_INTERFACE_IP_ADDRESS: 10.0.0.2 - EXTERNAL_INTERFACE: enp2s0 - NETWORK_CONFIG: enp2s0:dhcp;enp0s26u1u2:static,address=10.0.0.2,netmask=255.255.255.0 - SYNC_TIME_WITH_CONTROLLER: false - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - controller-to-install-initramfs: *initramfs -- morph: systems/installer-system-x86_64.morph - deploy: - compute-installer: - <<: *installer - location: installer-openstack-compute-x86_64.img - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - compute-initramfs: *initramfs - - morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - compute-to-install: - <<: *stack-node - HOSTNAME: threenode-compute - - NOVA_ENABLE_COMPUTE: true - - CINDER_ENABLE_CONTROLLER: false - CINDER_ENABLE_COMPUTE: true - CINDER_ENABLE_STORAGE: true - - NEUTRON_ENABLE_AGENT: true - NEUTRON_ENABLE_MANAGER: false - NEUTRON_ENABLE_CONTROLLER: false - - CEILOMETER_ENABLE_CONTROLLER: false - CEILOMETER_ENABLE_COMPUTE: true - - MANAGEMENT_INTERFACE_IP_ADDRESS: 10.0.0.3 - EXTERNAL_INTERFACE: eno1 - NETWORK_CONFIG: eno1:dhcp;enp0s29u1u3:static,address=10.0.0.3,netmask=255.255.255.0 - SYNC_TIME_WITH_CONTROLLER: true - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - compute-to-install-initramfs: *initramfs diff --git a/unmaintained/clusters/openstack-two-node-installer.morph b/unmaintained/clusters/openstack-two-node-installer.morph deleted file mode 100644 index e75b5a5e..00000000 --- a/unmaintained/clusters/openstack-two-node-installer.morph +++ /dev/null @@ -1,199 +0,0 @@ -name: openstack-two-node-installer -kind: cluster -description: | - This is a cluster morphology for deploying an installer for an x86_64 - OpenStack system spread across three nodes. - - This cluster creates disk images that may be `dd`'d onto install media to - produce an OpenStack cluster when instantiated. - - Alternatively it may be used to install directly onto a physical disk by - running: - - morph deploy clusters/openstack-two-node-installer.morph \ - controller-installer controller-installer.location=/dev/vdb - - Substituting contrller-installer for compute-installer will produce - different configurations, and it is possible to substitue /dev/vdb for a - different path to a disk image to install to a different disk image. - - Substitute the values of HOSTNAME, NETWORK_CONFIG, EXTERNAL_INTERFACE, - MANAGEMENT_IP_ADDRESS, CONTROLLER_HOST_ADDRESS, RABBITMQ_HOST and HOSTS_* to - match your hardware and networking configuration. - - Requirements to be able to run and test the system: - - - DISK_SIZE should be bigger than 5G - - The system has to have available at least 4G of RAM, but once - you start instantiating VMs you will need more. - - The IP of the system can't change, and you need to know it beforehand, - that is, the system needs a static IP address. - - This cluster is configurable, but with the following constraints: - - - The hostname in RABBITMQ_HOST has to match CONTROLLER_HOST_ADDRESS, - and HOST_CONTROLLER. - - HOSTS_CONTROLLER is only needed if the hostname (see previous point) - is not a FQDN. - - The IP listed in MANAGEMENT_INTERFACE_IP_ADDRESS has to match the one - used in HOSTS_CONTROLLER. - - CINDER_DEVICE should be a path to a storage device ready to be - used/formated for cinder data. - - EXTERNAL_INTERFACE is required when the system has more than one network - interface. - - You can also have a look at the following suggestions: - - - NOVA_VIRT_TYPE can be either 'kvm' or 'qemu', depending on where the - system is being deployed to. - - We recommend changing all the PASSWORDs variables, also the - KEYSTONE_TEMPORARY_ADMIN_TOKEN, METADATA_PROXY_SHARED_SECRET and - METERING_SECRET. - - Setting NOVA_BAREMETAL_SCHEDULING with a YAML truth value will configure - Nova to schedule baremetal machines through the Ironic driver, instead of - sheduling virtual machines. - -systems: -- morph: systems/installer-system-x86_64.morph - deploy: - controller-installer: &installer - type: extensions/rawdisk - location: installer-openstack-controller-x86_64.img - KERNEL_ARGS: init=/usr/lib/baserock-installer/installer - DISK_SIZE: 6G - HOSTNAME: installer-x86_64 - INSTALLER_TARGET_STORAGE_DEVICE: /dev/sda - INSTALLER_ROOTFS_TO_INSTALL: /rootfs - INSTALLER_POST_INSTALL_COMMAND: sync; poweroff -f - INITRAMFS_PATH: boot/initramfs.gz - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - controller-initramfs: &initramfs - type: extensions/initramfs - location: boot/initramfs.gz - - morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - controller-to-install: &stack-node - type: extensions/sysroot - location: rootfs - INSTALL_FILES: install-files/openstack/manifest - INITRAMFS_PATH: boot/initramfs.gz - - HOSTNAME: twonode-controller - - RABBITMQ_HOST: twonode-controller.os-mgmt - RABBITMQ_PORT: 5672 - RABBITMQ_USER: rabbitmq - RABBITMQ_PASSWORD: veryinsecure - - # This token needs to be unique and secret - KEYSTONE_ENABLE_SERVICE: true - KEYSTONE_TEMPORARY_ADMIN_TOKEN: 22f3aa1cf538e3f6d5e8 - KEYSTONE_ADMIN_PASSWORD: veryinsecure - KEYSTONE_DB_USER: keystoneDB - KEYSTONE_DB_PASSWORD: veryinsecure - - GLANCE_ENABLE_SERVICE: true - GLANCE_SERVICE_USER: glance - GLANCE_SERVICE_PASSWORD: veryinsecure - GLANCE_DB_USER: glanceDB - GLANCE_DB_PASSWORD: veryinsecure - - NOVA_ENABLE_CONTROLLER: true - NOVA_ENABLE_COMPUTE: false - NOVA_SERVICE_USER: nova - NOVA_SERVICE_PASSWORD: veryinsecure - NOVA_DB_USER: novaDB - NOVA_DB_PASSWORD: veryinsecure - NOVA_VIRT_TYPE: kvm - NOVA_BAREMETAL_SCHEDULING: false - - CINDER_ENABLE_CONTROLLER: true - CINDER_ENABLE_COMPUTE: false - CINDER_ENABLE_STORAGE: false - CINDER_SERVICE_USER: cinder - CINDER_SERVICE_PASSWORD: veryinsecure - CINDER_DB_USER: cinderDB - CINDER_DB_PASSWORD: veryinsecure - # Storage device to be used by Cinder - CINDER_DEVICE: /dev/sdb - - NEUTRON_ENABLE_AGENT: false - NEUTRON_ENABLE_MANAGER: true - NEUTRON_ENABLE_CONTROLLER: true - NEUTRON_SERVICE_USER: neutron - NEUTRON_SERVICE_PASSWORD: veryinsecure - NEUTRON_DB_USER: neutronDB - NEUTRON_DB_PASSWORD: veryinsecure - METADATA_PROXY_SHARED_SECRET: novaneutronmetasecret - - IRONIC_ENABLE_SERVICE: true - IRONIC_SERVICE_USER: ironic - IRONIC_SERVICE_PASSWORD: veryinsecure - IRONIC_DB_USER: ironicDB - IRONIC_DB_PASSWORD: veryinsecure - - CEILOMETER_SERVICE_USER: ceilometer - CEILOMETER_SERVICE_PASSWORD: veryinsecure - CEILOMETER_DB_USER: ceilometerDB - CEILOMETER_DB_PASSWORD: veryinsecure - CEILOMETER_ENABLE_CONTROLLER: true - CEILOMETER_ENABLE_COMPUTE: false - METERING_SECRET: insecureceilometersecret - - CONTROLLER_HOST_ADDRESS: twonode-controller.os-mgmt - MANAGEMENT_INTERFACE_IP_ADDRESS: 10.0.0.1 - - HOSTS_CONTROL: 10.0.0.1 twonode-controller.os-mgmt - HOSTS_COMPUTE: 10.0.0.3 twonode-compute.os-mgmt - EXTERNAL_INTERFACE: enp3s0 - NETWORK_CONFIG: enp3s0:dhcp;enp2s0:static,address=10.0.0.1,netmask=255.255.255.0 - SYNC_TIME_WITH_CONTROLLER: false - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - controller-to-install-initramfs: *initramfs -- morph: systems/installer-system-x86_64.morph - deploy: - compute-installer: - <<: *installer - location: installer-openstack-compute-x86_64.img - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - compute-initramfs: *initramfs - - morph: unmaintained/systems/openstack-system-x86_64.morph - deploy: - compute-to-install: - <<: *stack-node - HOSTNAME: twonode-compute - - KEYSTONE_ENABLE_SERVICE: false - - GLANCE_ENABLE_SERVICE: false - - NOVA_ENABLE_COMPUTE: true - NOVA_ENABLE_CONTROLLER: false - - CINDER_ENABLE_CONTROLLER: false - CINDER_ENABLE_COMPUTE: true - CINDER_ENABLE_STORAGE: true - - NEUTRON_ENABLE_AGENT: true - NEUTRON_ENABLE_MANAGER: false - NEUTRON_ENABLE_CONTROLLER: false - - IRONIC_ENABLE_SERVICE: false - - CEILOMETER_ENABLE_CONTROLLER: false - CEILOMETER_ENABLE_COMPUTE: true - - MANAGEMENT_INTERFACE_IP_ADDRESS: 10.0.0.3 - EXTERNAL_INTERFACE: eno1 - NETWORK_CONFIG: eno1:dhcp;enp0s29u1u3:static,address=10.0.0.3,netmask=255.255.255.0 - SYNC_TIME_WITH_CONTROLLER: true - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - compute-to-install-initramfs: *initramfs diff --git a/unmaintained/clusters/sdk-example-cluster.morph b/unmaintained/clusters/sdk-example-cluster.morph deleted file mode 100644 index 35d47508..00000000 --- a/unmaintained/clusters/sdk-example-cluster.morph +++ /dev/null @@ -1,46 +0,0 @@ -name: sdk-example-cluster -kind: cluster -description: | - An example of creating a cross-compile SDK for an embedded Baserock system. - - This cluster demonstrates how you can use the 'sdk' write extension to - produce a cross-compile SDK tarball for an Baserock applicance. In this - example the system is assumed to run on ARMv7, and the SDK is built to - run on any x86_32 GNU/Linux system. - - The SDK is a Baserock system itself, containing just 'build-essential' and a - 'cross-toolchain' stratum. The SDK system also includes the target - appliance's system, as a 'subsystem', so that the libraries and headers are - available when building. - - This cluster deploys the SDK system using the 'sdk' write extension, which - produces a tarball with a small shell header. When the shell header is - executed, and passed a directory name on the commandline, it extracts the SDK - to that path and patches the binaries so that they execute correctly from - that directory. - - Deploying the applicate system artifact to the target device should be - done with a separate cluster morphology, because you will often want to - do this without rebuilding the SDK. - - You must build each system with `morph build` before deploying. We recommend - doing this all from your Baserock development machine, using a Baserock - ARM distributed build network to produce the system artifact. Once both - system artifacts are cached locally, the `morph deploy` command will produce - a self-extracting shell script/tarball following the 'location' field. - - See the documentation of the sdk.write extension for more information. -systems: -- morph: systems/armv7lhf-cross-toolchain-system-x86_64.morph - deploy: - sdk: - type: extensions/sdk - location: armv7lhf-cross-toolchain-system-x86_64.sh - PREFIX: /usr - TARGET: armv7lhf-baserock-linux-gnueabi - subsystems: - - morph: unmaintained/systems/devel-system-armv7lhf-highbank.morph - deploy: - sysroot: - type: extensions/sysroot - location: usr/armv7lhf-baserock-linux-gnueabi/sys-root diff --git a/unmaintained/clusters/upgrade-devel.morph b/unmaintained/clusters/upgrade-devel.morph deleted file mode 100644 index 44759ac2..00000000 --- a/unmaintained/clusters/upgrade-devel.morph +++ /dev/null @@ -1,39 +0,0 @@ -name: upgrade-devel -kind: cluster -description: | - This is a cluster morphology that can be used to deploy systems to a - an existing Baserock devel system, as an upgrade of the running system. - - This method is for users who deployed a system manually from one of the - images provided on http://download.baserock.org. IT IS ONLY POSSIBLE TO - UPGRADE BASEROCK 14 RELEASES OR NEWER. - - If you deployed your system using `morph deploy` then you should reuse the - cluster morphology you did the initial deployment with, instead of this one, - so that the configuration is preserved in the new system. - - Ensure that your root user has passwordless SSH access to localhost with - `ssh root@localhost whoami`. If not, run `ssh-copy-id root@localhost`. - Make sure the 'morph' field below matches the system you are upgrading. - - To upgrade, select a sensible a value for VERSION_LABEL and run: - - morph upgrade clusters/upgrade-devel.morph self.HOSTNAME=$(hostname) self.VERSION_LABEL=$VERSION_LABEL - - Your configuration in /etc should be propagated to the new system, but there - may be merge conflicts. Check /etc for files named '.rej' and '.orig' in the - new system, which will indicate that there are changes from the old system - that must be merged manually. You can get a nice diff from the old /etc as - follows: - - mount /dev/sda /mnt - git diff --no-index /mnt/systems/factory/run/etc /mnt/systems/$VERSION_LABEL/run/etc - - On a base system, use 'diff -r' instead of 'git diff --no-index'. It will - not be as colourful. -systems: -- morph: unmaintained/systems/devel-system-x86_64-generic.morph - deploy: - self: - upgrade-type: extensions/ssh-rsync - upgrade-location: root@127.0.0.1 diff --git a/unmaintained/clusters/xfce-system-x86_64-deploy.morph b/unmaintained/clusters/xfce-system-x86_64-deploy.morph deleted file mode 100644 index 465ba1d0..00000000 --- a/unmaintained/clusters/xfce-system-x86_64-deploy.morph +++ /dev/null @@ -1,23 +0,0 @@ -name: xfce-system-x86_64-deploy -kind: cluster -description: | - Deploy a stock XFCE system. - - The resulting image can be copied to a USB and booted from there, - as well as in a virtual machine. - -systems: -- morph: unmaintained/systems/xfce-system-x86_64.morph - deploy: - xfce-system-x86_64: - type: extensions/rawdisk - location: /xfce-system-x86_64.img - DISK_SIZE: 6G - KERNEL_ARGS: vga=788 - INITRAMFS_PATH: boot/initramfs.gz - subsystems: - - morph: systems/initramfs-x86_64.morph - deploy: - initramfs: - type: extensions/initramfs - location: boot/initramfs.gz diff --git a/unmaintained/clusters/zookeeper.morph b/unmaintained/clusters/zookeeper.morph deleted file mode 100644 index ba859bc1..00000000 --- a/unmaintained/clusters/zookeeper.morph +++ /dev/null @@ -1,21 +0,0 @@ -name: zookeeper -kind: cluster -systems: -- morph: unmaintained/systems/zookeeper-client-x86_64.morph - deploy: - my-client-system: - type: extensions/kvm - location: kvm+ssh://username@HOSTNAME/machinename/path/to/zookeeper-client.img - DISK_SIZE: 4G - RAM_SIZE: 1G - VCPUS: 1 - HOSTNAME: zkclient -- morph: unmaintained/systems/zookeeper-server-x86_64.morph - deploy: - my-server-system: - type: extensions/kvm - location: kvm+ssh://username@HOSTNAME/machinename/path/to/zookeeper-server.img - DISK_SIZE: 4G - RAM_SIZE: 1G - VCPUS: 1 - HOSTNAME: zkserver |