Version 9 (modified by 12 years ago) (diff) | ,
---|
DETER Recommended Hardware
We have a some recommended hardware for setting up DETER. We can't hand you a shopping list, but here are the things that we use and like:
Network Switches
The DETER software stack dynamically sets up real VLANs on switches. There are also two networks in a typical DETER/Emulab install. The control network is where testbed nodes network boot and mount filesystems over. The experimental network is where experimental topologies are instantiated. These networks may be separated or coexists on the same switch. It depends on how big your testbed is and if you mind control/fileserver traffic going over the same switch trunks with experimental data. For small installs, a single switch is fine.
Although you may be tempted to borrow an engineering sample PacketMasher 9000 from that cool startup your friend works at in Silicon Valley, it is important that you choose a switch that has software support. Adding support for new switches is possible, but non-trivial. Right now, we recommend HP 5400zl series switches. The number of ports depends on how many nodes you want to support. At DETER, we typically have at least five interfaces per testbed node.
For really small installations, we have had good luck with HP 2810 switches:
Infrastructure Machines
There are three main machines in a DETER/Emulab installation. These are boss, users, and router. Boss hosts the database, web interface, and main logic. Users acts as the NFS/SMB server and user login machine. All these machines run FreeBSD 9.1.
These machines do not need to be very high powered if your budget is limited. We have successfully deployed all three machines on a single PowerEdge 860 with 4GB of ram running VMWare ESXi. How much of a box your provision is really up to what your site requirements are.
Testbed nodes
The Emulab software stack that powers DETER predates the 'Cloud' buzzword by about a decade. Although we support a virtualization overlay through DETER Containers, we are really about handing researchers physical machines and physical networks. At DETER we tend to purchase low end, single CPU server class machines. Things like advanced remote management, redundant power supplies, and over engineered hardware are best left to the infrastructure machines described above. If you have a choice between doing 16 entry level machines or 4 over powered nodes, we currently recommend going for quantity.
What you provision here is really up to you. All of the machines we are using at DETER are no longer in production, so we can not recommend specific models.
- The machines must be able to PXE boot.
- The machines need to be capable of running FreeBSD 8.3. Watch out for fancy RAID controllers or strange network cards. Standard SATA drives and Intel NICs are what we use.
- You should ideally have about 5 Gbe network interfaces on each node. One interface for the control network and four for the experimental network.
For our most recent buildout, we used SuperMicro MicroCloud machines. We get 8 Xeon E3 servers in 3u of rack space and control them via IPMI.
Aside from these basic requirements, the testbed nodes depend on what you intend to do.
Our standard operating system images include Ubuntu 12.04, CentOS 6, and FreeBSD 9.
Power and Serial Controllers
Historically, we have used real power controllers and serial concentrators rather than IPMI. We are also currently using IPMI, so you can skip on the power and serial controllers if you want. So far we have been successful using the SuperMicro Nuvoton WPCM450RA0BX found on their MicroCloud systems. Adding support for other IPMI controllers should not be too bad if they can be made to work with open source IPMI tools that compile on FreeBSD.
For power controllers, we are using APC 7902 rack PDUs.
For serial controllers, we are using Digi Etherlite models.