The test environment is build on Cisco UCS B-200 M2 blades in a Cisco Blade chassis with dual UCS 6100 switches and Cisco VIC adapters. The example site.pp uses the Flat_DHCP model, and assumes that an eth0 interface will be attached to the "public" newtork, and that an eth1 interface will be avaialble for the Flat network bridge.
The blade config is:
2x 6 core processors 96GB memory Cisco VIC 2x 300GB SATA drives in a Raid-1 configuration
The system profile has two VNICs, one on VLAN 100 (arbitrary), and one on VLAN 105 (also arbitrary). VLAN 100 is internet accessible via a NAT gateway and via inbound VPN via an ASA firewall. VLAN 105 is unconnected
"Public" addresses on VLAN100 are in the 192.168.100.0/24 range, and a static pool (nova pool) is carved out of that range as 192.168.100.64/28 "Private" addresses are from the floating pool in the 10.0.0.0/16 range, but a network in the 10.0.0.0/24 range is actaully assigned to the openstack tenant.
Building the environment
Build an Ubuntu 12.04 system.
We still need a build node (though will plan to migrate to a RAMFS based system soon), and you can use the preseed file in the example directory to build the base os. Start with the ISO boot (or USB boot), and at the initial installer screen, after you pick your language, hit F6 (or FN-F6 on a mac), and ESC, and add (don't delete anything from the current boot string):
If you can't get to the 128 address (i.e., you're outside of Cisco), host the preseed file on your own machine. You may need to add network information (unless you have DHCP enabled, which you may want to disable and give control over to cobbler).
If you don't want to go the pre-seed route, then build a machine normally, make sure you have ~10GB of disk space available, and I recommend using an LVM volume rather than just using the whole disk, but it's less critical for this build.
Once the node is built log in (localadmin:ubuntu are the default), and become root (I usually do "sudo -H bash"), or preface all the following commands with "sudo".
you will still need to log into the console of the control node to load in an image: user: localadmin, password: ubuntu. If you SU to root, there is an openrc auth file in root's home directory, and you can launch a test file in /tmp/nova_test.sh.
You should now have a cirros image and a running instance (called dans_vm if you didn't change anything).
Adding in Swift
Swift adds a scaleable redundant object storage system to the openstack environment, and is a core part of the general model. In order to add swift, you will need a minimum of 3 additional machines, prefereably with a large number of fast hard drives, with RAID preferably disabled (that's right RAID is not recommended for Swift clusters).
Since the system does not yet catalog devices, you will need to manually define the devices to be used by the swift system and configure the appropriate "nodes" and "rings" for the number of devices defined. The included example is set up for two devices, either LVM based devices (useful if you are using UCS blades which nominally only support a maximum of 2 drives today or are working in a constrained development environment) or perferably raw direct access devices. The puppet modules will attempt to format the entire device as an XFS file system, and then add them in to the system.
EDIT THE swift-nodes.pp FILE before including it in your site.pp file. Then you can add the swift-nodes.pp file into your site.pp (or uncomment it from the example file):
Build/re-build the swift machines, and you should have a swift capable system.