Changes between Version 12 and Version 13 of GENIRacksHome


Ignore:
Timestamp:
03/07/12 10:01:01 (8 years ago)
Author:
lnevers@bbn.com
Comment:

--

Legend:

Unmodified
Added
Removed
Modified
  • GENIRacksHome

    v12 v13  
    55GENI racks are being implemented to meet the goals specified by the GPO [wiki:GeniRacks GENI Rack Requirements].  Current rack projects can be characterized as follows:
    66
     7
     8 * '''[#ExoGENIRacks ExoGENI]'''  A higher cost, flexible virtual networking topologies solution including !OpenFlow, that also delivers a powerful platform for multi-site cloud applications. These racks are typically deployed as an integrated part of a campus network.
     9 * '''[#InstaGENIRacks InstaGENI]''' - A mid-range cost, expandable GENI Racks solution that can will be deployed at a large number of campuses, delivering Internet cloud applications support, along with !Openflow and VLAN networking. These racks are normally deployed outside a site firewall.
    710 * '''[#GENIStarterRacks Starter Racks]''': Deployed by the GPO, a project to deliver an early low-end solution for GENI Racks.
    8  * '''[#InstaGENIRacks InstaGENI]''' - A mid-range cost, expandable GENI Racks solution that can will be deployed at a large number of campuses, delivering Internet cloud applications support, along with !Openflow and VLAN networking. These racks are normally deployed outside a site firewall.
    9  * '''[#ExoGENIRacks ExoGENI]'''  A higher cost, flexible virtual networking topologies solution including !OpenFlow, that also delivers a powerful platform for multi-site cloud applications. These racks are typically deployed as an integrated part of a campus network.
     11
     12The GPO recommends the ExoGENI or InstaGENI Racks for broad campus deployment.
    1013
    1114All GENI Racks have layer 3 connections to the internet and layer 2 connections to the GENI core networks (curently NLR and Internet2). The racks use commodity Internet for control access to rack resources, and shared VLANS for the application and experiment data connections.  The racks may also use layer 3 Internet connections for some experiments, particularly IP cloud experiments.  This diagram illustrates logical connections for all the GENI Rack types:
     
    1720'''[wiki:GENIRacksAdministration GENI Racks Administration]''' tasks are highlighted for each of these rack solutions to provide some insight into the effort required by a participant.
    1821
     22
     23
     24= ExoGENI Racks =
     25
     26The [http://www.exogeni.net ExoGENI] project is an effort to implement high-performance GENI Racks via a partnership between RENaissance Computing Institute (RENCI), Duke and IBM.  ExoGENI racks are assembled and tested by IBM and shipped directly to sites, where they are managed by the RENCI team.  ExoGENI racks deliver support for multi-domain cloud structure with flexible virtual networking topologies that allow combining ExoGENI, Meso-scale !OpenFlow and WiMAX resource.  An overview of this project was presented at the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/GEC12-ExoGENI-Racks.pptx?format=raw GEC12 ExoGENI Presentation]. Also available are an [https://code.renci.org/gf/download/docmanfileversion/13/691/ExoGENIDesignv1.02.pdf ExoGENI Design] document and an [http://www.cs.duke.edu/~chase/exogeni.pdf ExoGENI white paper].
     27
     28== ExoGENI Components ==
     29
     30An ExoGENI Rack delivers the following types of systems:
     31 * Management Switch - An IBM G8052R switched is delivered to allow access to/from the local network provider.
     32 * VPN appliance - A Juniper SSG5 provides backup access to manage nodes.
     33 * !OpenFlow-enabled switch - An IBM G8264R switch to carry experimental traffic via layer2 to GENI backbone (I2, NLR) and to local OF campus.
     34 * Management Node - An IBM x3650 host running Elastic Compute Cloud(EC2) with !OpenStack to provision VMs and running xCat to provision bare-metal nodes. Also runs monitoring functions.
     35 * Worker Nodes - Ten IBM 3650 M3 Worker nodes provide both !OpenStack virtualized instances and Bare-metal Xcat nodes
     36 * Monitoring Host - None, Monitoring is through Nagios from GMOC.
     37
     38The ExoGENI resources have the following connections:
     39
     40[[Image(GENI-ExoGENI-components.jpg)]]
     41
     42== ExoGENI Specifications ==
     43
     44An initial inventory of the ExoGENI Rack hardware components is found [https://docs.google.com/document/d/1hzleT6TNmiDb0YkkgqjXxPFJ37P4O6qApLmXgKJHBZQ/edit?hl=en_US here], which is superseded by the following:
     45
     46''__Compute Resource__'' A total of 12 hosts are in the rack to provide the Resources, Monitoring, Storage and Application functions:
     47  * Management node: 1 IBM x3650 M3, 2x146GB 10K SAS hard drives, 12G RAM, dual-socket 4-core Intel X5650 2.66Ghz CPU, Quad-port 1Gbps adapter
     48  * Worker/Bare-Metal nodes: 10 IBM x3650 M3, 1x146GB 10K SAS hard drive +1x500+GB secondary drive, 48G RAM, dual-socket 6-core Intel X5650 2.66Ghz CPU, dual 1Gbps adapter, 10G dual-port Chelseo adapter
     49  * Sliverable Storage: 1 IBM DS3512 storage NAS 6x1TB 7200RPM drives
     50
     51''__Network Components__''
     52 * Management Switch: IBM BNT G8052R 1G client/10G uplink ports - Access to commodity internet.
     53 * !OpenFlow Switch: IBM BNT G8264R 10G client/40G uplink ports - Access to GENI backbone.
     54 * VPN Appliance: Juniper SSG5 - Backup management access.
     55
     56''__Misc. Components__'' General purpose hardware included:
     57  * IBM PDU based on site power requirements, (GPO=IBM 5897 PDU; RENCI=DPI 5900 PDU)
     58  * No UPS included
     59  * IBM Local 2x16 Console Manager (LCM16)
     60  * IBM 1U 17-inch Flat Panel Console Kit (PN 172317X)
     61
     62== ExoGENI Implementation ==
     63Any configuration or run-time requirements for the ExoGENI rack systems will be documented here, when available.
     64
     65Each ExoGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, Management node, Worker Nodes, etc.. See the '''[wiki:GENIRacks/ExoGENICompSetup ExoGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#ExoGENIAdministration ExoGENI Racks Administration]''' tasks.
     66
     67== ExoGENI Monitoring ==
     68Monitoring data for the ExoGENI rack is collected on the management node by a Nagios aggregator and then forwarded to the GMOC. The type of data that will be available is currently being defined.  ExoGENI Racks meet the GENI [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. ExoGENI Monitoring data will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP Home].
     69
     70== ExoGENI Site Requirements ==
     71ExoGENI racks have the following site requirements:
     72 * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc.
     73 * Site Security Requirements- Determine changes required for rack connectivity, such as FV rules, IP filters, etc.
     74 * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components.
     75 * Power requirements - based on site requirements
     76 * Administrative accounts - Setup of site administrator account that will be created on the management/head node.
     77 * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place.  Also covers any physical restrictions for the rack delivery.
     78
     79If you are interested in becoming an ExoGENI deployment site, please contact us at [mailto:ibaldin@renci.org].
     80------
     81= InstaGENI Racks =
     82The InstaGENI project is an effort to deploy low-end, expandable GENI Racks at large number of campuses and is typically found outside the site firewall, more details are to be added as defined. An overview of the project is available from the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/McGeer%20InstaGENI.pdf?format=raw  GEC12 InstaGENI Racks] presentation. An [http://groups.geni.net/geni/attachment/wiki/GENIRacksHome/InstaGENI%20Design%20Document%20Rick%202-15%282%29.doc?format=raw InstaGENI Design] document is available.
     83
     84== InstaGENI Components ==
     85
     86Each InstaGENI racks delivers a small ProtoGENI cluster with !OpenFlow networking and FOAM aggregate management. This rack includes the following types of systems:
     87   * Control Node -  Xen Server that runs 3 VMS to provide:
     88            * PG boss node, web server, monitoring, and GENI API Server
     89            * Local File Server node
     90            * FOAM Controller
     91   * Experiment Nodes - Five nodes managed by ProtoGENI software stack, which provides boot services, account creation, experimental management, etc.
     92   * !OpenFlow Switch - Provides internal routing and data plane connectivity to the GENI backbone (layer 2 and layer 3)
     93   * Management Switch - Provides control plane connectivity to the Internet (layer 3)
     94
     95Following are the network connections for an InstaGENI rack:
     96
     97[[Image(GENI-InstaGENI-components.jpg)]]
     98
     99== InstaGENI Specifications ==
     100The current hardware components specification for the InstaGENI Racks includes:
     101
     102''__Compute Resource Specifications__'' 5 HP !ProLiant DL360 G7 Server series hosts to provide the VM Server, Monitoring, Storage, and Application functions:
     103  * Control Node - 1 HP !ProLiant DL360 G7 Server, quad-core, single-socket, 12 GB Ram, 4 TB Disk (RAID), and dual NIC
     104  * Experiment Nodes - 5 HP !ProLiant DL360 G7 Server, six-core, dual-socket,48GB Ram, 1TB Disk, and dual NIC
     105  * Bare Metal Node
     106
     107''__Network Components__'' X network components to provide access to GENI Backbone and Commodity Internet:
     108 * HP !ProCurve 2610 switch, 24 10/100 Mb/s ports, 2 1 Gb/s ports
     109 * HP !ProCurve 6600 switch 48 1 Gb/s ports, 4 10 Gb/s ports
     110
     111''__Misc. Components_'' General purpose hardware also included:
     112  * 1 or more HP Intelignt Mod PDU 
     113  * HP Dedicated iLO Management Port Option
     114  * HP TFT7600 KVM Console US Kit
     115
     116== InstaGENI Implementation ==
     117Any configuration or run-time requirements for any of the InstaGENI rack systems will be documented here, when available.
     118
     119Each InstaGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, control node, experiment nodes, bare-metal nodes, etc.. See the '''[wiki:GENIRacks/InstaGENICompSetup InstaGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#InstaGENIAdministration InstaGENIRacks Administration]''' tasks.
     120
     121== InstaGENI Monitoring ==
     122InstaGENI Racks meet the following [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements].  InstaGENI Monitoring data is currently being defined and will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP home].
     123
     124== InstaGENI Site Requirements ==
     125InstaGENI racks have the following site requirements:
     126 * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc.
     127 * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components.
     128 * Power requirements - based on site requirements
     129 * Administrative accounts - Setup of site administrator account that will be created on the management/head node.
     130 * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place.  Also covers any physical restrictions for the rack delivery.
     131
     132If you are interested in becoming an InstaGENI deployment site, please contact us at [mailto:rick.mcgeer@hp.com].
     133
     134------
    19135= GENI Starter Racks =
    20136
     
    73189------
    74190
    75 = InstaGENI Racks =
    76 The InstaGENI project is an effort to deploy low-end, expandable GENI Racks at large number of campuses and is typically found outside the site firewall, more details are to be added as defined. An overview of the project is available from the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/McGeer%20InstaGENI.pdf?format=raw  GEC12 InstaGENI Racks] presentation. An [http://groups.geni.net/geni/attachment/wiki/GENIRacksHome/InstaGENI%20Design%20Document%20Rick%202-15%282%29.doc?format=raw InstaGENI Design] document is available.
    77 
    78 == InstaGENI Components ==
    79 
    80 Each InstaGENI racks delivers a small ProtoGENI cluster with !OpenFlow networking and FOAM aggregate management. This rack includes the following types of systems:
    81    * Control Node -  Xen Server that runs 3 VMS to provide:
    82             * PG boss node, web server, monitoring, and GENI API Server
    83             * Local File Server node
    84             * FOAM Controller
    85    * Experiment Nodes - Five nodes managed by ProtoGENI software stack, which provides boot services, account creation, experimental management, etc.
    86    * !OpenFlow Switch - Provides internal routing and data plane connectivity to the GENI backbone (layer 2 and layer 3)
    87    * Management Switch - Provides control plane connectivity to the Internet (layer 3)
    88 
    89 Following are the network connections for an InstaGENI rack:
    90 
    91 [[Image(GENI-InstaGENI-components.jpg)]]
    92 
    93 == InstaGENI Specifications ==
    94 The current hardware components specification for the InstaGENI Racks includes:
    95 
    96 ''__Compute Resource Specifications__'' 5 HP !ProLiant DL360 G7 Server series hosts to provide the VM Server, Monitoring, Storage, and Application functions:
    97   * Control Node - 1 HP !ProLiant DL360 G7 Server, quad-core, single-socket, 12 GB Ram, 4 TB Disk (RAID), and dual NIC
    98   * Experiment Nodes - 5 HP !ProLiant DL360 G7 Server, six-core, dual-socket,48GB Ram, 1TB Disk, and dual NIC
    99   * Bare Metal Node
    100 
    101 ''__Network Components__'' X network components to provide access to GENI Backbone and Commodity Internet:
    102  * HP !ProCurve 2610 switch, 24 10/100 Mb/s ports, 2 1 Gb/s ports
    103  * HP !ProCurve 6600 switch 48 1 Gb/s ports, 4 10 Gb/s ports
    104 
    105 ''__Misc. Components_'' General purpose hardware also included:
    106   * 1 or more HP Intelignt Mod PDU 
    107   * HP Dedicated iLO Management Port Option
    108   * HP TFT7600 KVM Console US Kit
    109 
    110 == InstaGENI Implementation ==
    111 Any configuration or run-time requirements for any of the InstaGENI rack systems will be documented here, when available.
    112 
    113 Each InstaGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, control node, experiment nodes, bare-metal nodes, etc.. See the '''[wiki:GENIRacks/InstaGENICompSetup InstaGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#InstaGENIAdministration InstaGENIRacks Administration]''' tasks.
    114 
    115 == InstaGENI Monitoring ==
    116 InstaGENI Racks meet the following [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements].  InstaGENI Monitoring data is currently being defined and will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP home].
    117 
    118 == InstaGENI Site Requirements ==
    119 InstaGENI racks have the following site requirements:
    120  * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc.
    121  * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components.
    122  * Power requirements - based on site requirements
    123  * Administrative accounts - Setup of site administrator account that will be created on the management/head node.
    124  * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place.  Also covers any physical restrictions for the rack delivery.
    125 
    126 If you are interested in becoming an InstaGENI deployment site, please contact us at [mailto:rick.mcgeer@hp.com].
    127 
    128 --------------------
    129 
    130 = ExoGENI Racks =
    131 
    132 The [http://www.exogeni.net ExoGENI] project is an effort to implement high-performance GENI Racks via a partnership between RENaissance Computing Institute (RENCI), Duke and IBM.  ExoGENI racks are assembled and tested by IBM and shipped directly to sites, where they are managed by the RENCI team.  ExoGENI racks deliver support for multi-domain cloud structure with flexible virtual networking topologies that allow combining ExoGENI, Meso-scale !OpenFlow and WiMAX resource.  An overview of this project was presented at the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/GEC12-ExoGENI-Racks.pptx?format=raw GEC12 ExoGENI Presentation]. Also available are an [https://code.renci.org/gf/download/docmanfileversion/13/691/ExoGENIDesignv1.02.pdf ExoGENI Design] document and an [http://www.cs.duke.edu/~chase/exogeni.pdf ExoGENI white paper].
    133 
    134 == ExoGENI Components ==
    135 
    136 An ExoGENI Rack delivers the following types of systems:
    137  * Management Switch - An IBM G8052R switched is delivered to allow access to/from the local network provider.
    138  * VPN appliance - A Juniper SSG5 provides backup access to manage nodes.
    139  * !OpenFlow-enabled switch - An IBM G8264R switch to carry experimental traffic via layer2 to GENI backbone (I2, NLR) and to local OF campus.
    140  * Management Node - An IBM x3650 host running Elastic Compute Cloud(EC2) with !OpenStack to provision VMs and running xCat to provision bare-metal nodes. Also runs monitoring functions.
    141  * Worker Nodes - Ten IBM 3650 M3 Worker nodes provide both !OpenStack virtualized instances and Bare-metal Xcat nodes
    142  * Monitoring Host - None, Monitoring is through Nagios from GMOC.
    143 
    144 The ExoGENI resources have the following connections:
    145 
    146 [[Image(GENI-ExoGENI-components.jpg)]]
    147 
    148 == ExoGENI Specifications ==
    149 
    150 An initial inventory of the ExoGENI Rack hardware components is found [https://docs.google.com/document/d/1hzleT6TNmiDb0YkkgqjXxPFJ37P4O6qApLmXgKJHBZQ/edit?hl=en_US here], which is superseded by the following:
    151 
    152 ''__Compute Resource__'' A total of 12 hosts are in the rack to provide the Resources, Monitoring, Storage and Application functions:
    153   * Management node: 1 IBM x3650 M3, 2x146GB 10K SAS hard drives, 12G RAM, dual-socket 4-core Intel X5650 2.66Ghz CPU, Quad-port 1Gbps adapter
    154   * Worker/Bare-Metal nodes: 10 IBM x3650 M3, 1x146GB 10K SAS hard drive +1x500+GB secondary drive, 48G RAM, dual-socket 6-core Intel X5650 2.66Ghz CPU, dual 1Gbps adapter, 10G dual-port Chelseo adapter
    155   * Sliverable Storage: 1 IBM DS3512 storage NAS 6x1TB 7200RPM drives
    156 
    157 ''__Network Components__''
    158  * Management Switch: IBM BNT G8052R 1G client/10G uplink ports - Access to commodity internet.
    159  * !OpenFlow Switch: IBM BNT G8264R 10G client/40G uplink ports - Access to GENI backbone.
    160  * VPN Appliance: Juniper SSG5 - Backup management access.
    161 
    162 ''__Misc. Components__'' General purpose hardware included:
    163   * IBM PDU based on site power requirements, (GPO=IBM 5897 PDU; RENCI=DPI 5900 PDU)
    164   * No UPS included
    165   * IBM Local 2x16 Console Manager (LCM16)
    166   * IBM 1U 17-inch Flat Panel Console Kit (PN 172317X)
    167 
    168 == ExoGENI Implementation ==
    169 Any configuration or run-time requirements for the ExoGENI rack systems will be documented here, when available.
    170 
    171 Each ExoGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, Management node, Worker Nodes, etc.. See the '''[wiki:GENIRacks/ExoGENICompSetup ExoGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#ExoGENIAdministration ExoGENI Racks Administration]''' tasks.
    172 
    173 == ExoGENI Monitoring ==
    174 Monitoring data for the ExoGENI rack is collected on the management node by a Nagios aggregator and then forwarded to the GMOC. The type of data that will be available is currently being defined.  ExoGENI Racks meet the GENI [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. ExoGENI Monitoring data will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP Home].
    175 
    176 == ExoGENI Site Requirements ==
    177 ExoGENI racks have the following site requirements:
    178  * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc.
    179  * Site Security Requirements- Determine changes required for rack connectivity, such as FV rules, IP filters, etc.
    180  * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components.
    181  * Power requirements - based on site requirements
    182  * Administrative accounts - Setup of site administrator account that will be created on the management/head node.
    183  * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place.  Also covers any physical restrictions for the rack delivery.
    184 
    185 If you are interested in becoming an ExoGENI deployment site, please contact us at [mailto:ibaldin@renci.org].
    186 -----
    187 
    188191= Get Your Own GENI Rack =
    189192If you are interested in making your own GENI Starter Rack, the GPO can help with the following: