| 22 | |
| 23 | |
| 24 | = ExoGENI Racks = |
| 25 | |
| 26 | The [http://www.exogeni.net ExoGENI] project is an effort to implement high-performance GENI Racks via a partnership between RENaissance Computing Institute (RENCI), Duke and IBM. ExoGENI racks are assembled and tested by IBM and shipped directly to sites, where they are managed by the RENCI team. ExoGENI racks deliver support for multi-domain cloud structure with flexible virtual networking topologies that allow combining ExoGENI, Meso-scale !OpenFlow and WiMAX resource. An overview of this project was presented at the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/GEC12-ExoGENI-Racks.pptx?format=raw GEC12 ExoGENI Presentation]. Also available are an [https://code.renci.org/gf/download/docmanfileversion/13/691/ExoGENIDesignv1.02.pdf ExoGENI Design] document and an [http://www.cs.duke.edu/~chase/exogeni.pdf ExoGENI white paper]. |
| 27 | |
| 28 | == ExoGENI Components == |
| 29 | |
| 30 | An ExoGENI Rack delivers the following types of systems: |
| 31 | * Management Switch - An IBM G8052R switched is delivered to allow access to/from the local network provider. |
| 32 | * VPN appliance - A Juniper SSG5 provides backup access to manage nodes. |
| 33 | * !OpenFlow-enabled switch - An IBM G8264R switch to carry experimental traffic via layer2 to GENI backbone (I2, NLR) and to local OF campus. |
| 34 | * Management Node - An IBM x3650 host running Elastic Compute Cloud(EC2) with !OpenStack to provision VMs and running xCat to provision bare-metal nodes. Also runs monitoring functions. |
| 35 | * Worker Nodes - Ten IBM 3650 M3 Worker nodes provide both !OpenStack virtualized instances and Bare-metal Xcat nodes |
| 36 | * Monitoring Host - None, Monitoring is through Nagios from GMOC. |
| 37 | |
| 38 | The ExoGENI resources have the following connections: |
| 39 | |
| 40 | [[Image(GENI-ExoGENI-components.jpg)]] |
| 41 | |
| 42 | == ExoGENI Specifications == |
| 43 | |
| 44 | An initial inventory of the ExoGENI Rack hardware components is found [https://docs.google.com/document/d/1hzleT6TNmiDb0YkkgqjXxPFJ37P4O6qApLmXgKJHBZQ/edit?hl=en_US here], which is superseded by the following: |
| 45 | |
| 46 | ''__Compute Resource__'' A total of 12 hosts are in the rack to provide the Resources, Monitoring, Storage and Application functions: |
| 47 | * Management node: 1 IBM x3650 M3, 2x146GB 10K SAS hard drives, 12G RAM, dual-socket 4-core Intel X5650 2.66Ghz CPU, Quad-port 1Gbps adapter |
| 48 | * Worker/Bare-Metal nodes: 10 IBM x3650 M3, 1x146GB 10K SAS hard drive +1x500+GB secondary drive, 48G RAM, dual-socket 6-core Intel X5650 2.66Ghz CPU, dual 1Gbps adapter, 10G dual-port Chelseo adapter |
| 49 | * Sliverable Storage: 1 IBM DS3512 storage NAS 6x1TB 7200RPM drives |
| 50 | |
| 51 | ''__Network Components__'' |
| 52 | * Management Switch: IBM BNT G8052R 1G client/10G uplink ports - Access to commodity internet. |
| 53 | * !OpenFlow Switch: IBM BNT G8264R 10G client/40G uplink ports - Access to GENI backbone. |
| 54 | * VPN Appliance: Juniper SSG5 - Backup management access. |
| 55 | |
| 56 | ''__Misc. Components__'' General purpose hardware included: |
| 57 | * IBM PDU based on site power requirements, (GPO=IBM 5897 PDU; RENCI=DPI 5900 PDU) |
| 58 | * No UPS included |
| 59 | * IBM Local 2x16 Console Manager (LCM16) |
| 60 | * IBM 1U 17-inch Flat Panel Console Kit (PN 172317X) |
| 61 | |
| 62 | == ExoGENI Implementation == |
| 63 | Any configuration or run-time requirements for the ExoGENI rack systems will be documented here, when available. |
| 64 | |
| 65 | Each ExoGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, Management node, Worker Nodes, etc.. See the '''[wiki:GENIRacks/ExoGENICompSetup ExoGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#ExoGENIAdministration ExoGENI Racks Administration]''' tasks. |
| 66 | |
| 67 | == ExoGENI Monitoring == |
| 68 | Monitoring data for the ExoGENI rack is collected on the management node by a Nagios aggregator and then forwarded to the GMOC. The type of data that will be available is currently being defined. ExoGENI Racks meet the GENI [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. ExoGENI Monitoring data will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP Home]. |
| 69 | |
| 70 | == ExoGENI Site Requirements == |
| 71 | ExoGENI racks have the following site requirements: |
| 72 | * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc. |
| 73 | * Site Security Requirements- Determine changes required for rack connectivity, such as FV rules, IP filters, etc. |
| 74 | * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components. |
| 75 | * Power requirements - based on site requirements |
| 76 | * Administrative accounts - Setup of site administrator account that will be created on the management/head node. |
| 77 | * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place. Also covers any physical restrictions for the rack delivery. |
| 78 | |
| 79 | If you are interested in becoming an ExoGENI deployment site, please contact us at [mailto:ibaldin@renci.org]. |
| 80 | ------ |
| 81 | = InstaGENI Racks = |
| 82 | The InstaGENI project is an effort to deploy low-end, expandable GENI Racks at large number of campuses and is typically found outside the site firewall, more details are to be added as defined. An overview of the project is available from the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/McGeer%20InstaGENI.pdf?format=raw GEC12 InstaGENI Racks] presentation. An [http://groups.geni.net/geni/attachment/wiki/GENIRacksHome/InstaGENI%20Design%20Document%20Rick%202-15%282%29.doc?format=raw InstaGENI Design] document is available. |
| 83 | |
| 84 | == InstaGENI Components == |
| 85 | |
| 86 | Each InstaGENI racks delivers a small ProtoGENI cluster with !OpenFlow networking and FOAM aggregate management. This rack includes the following types of systems: |
| 87 | * Control Node - Xen Server that runs 3 VMS to provide: |
| 88 | * PG boss node, web server, monitoring, and GENI API Server |
| 89 | * Local File Server node |
| 90 | * FOAM Controller |
| 91 | * Experiment Nodes - Five nodes managed by ProtoGENI software stack, which provides boot services, account creation, experimental management, etc. |
| 92 | * !OpenFlow Switch - Provides internal routing and data plane connectivity to the GENI backbone (layer 2 and layer 3) |
| 93 | * Management Switch - Provides control plane connectivity to the Internet (layer 3) |
| 94 | |
| 95 | Following are the network connections for an InstaGENI rack: |
| 96 | |
| 97 | [[Image(GENI-InstaGENI-components.jpg)]] |
| 98 | |
| 99 | == InstaGENI Specifications == |
| 100 | The current hardware components specification for the InstaGENI Racks includes: |
| 101 | |
| 102 | ''__Compute Resource Specifications__'' 5 HP !ProLiant DL360 G7 Server series hosts to provide the VM Server, Monitoring, Storage, and Application functions: |
| 103 | * Control Node - 1 HP !ProLiant DL360 G7 Server, quad-core, single-socket, 12 GB Ram, 4 TB Disk (RAID), and dual NIC |
| 104 | * Experiment Nodes - 5 HP !ProLiant DL360 G7 Server, six-core, dual-socket,48GB Ram, 1TB Disk, and dual NIC |
| 105 | * Bare Metal Node |
| 106 | |
| 107 | ''__Network Components__'' X network components to provide access to GENI Backbone and Commodity Internet: |
| 108 | * HP !ProCurve 2610 switch, 24 10/100 Mb/s ports, 2 1 Gb/s ports |
| 109 | * HP !ProCurve 6600 switch 48 1 Gb/s ports, 4 10 Gb/s ports |
| 110 | |
| 111 | ''__Misc. Components_'' General purpose hardware also included: |
| 112 | * 1 or more HP Intelignt Mod PDU |
| 113 | * HP Dedicated iLO Management Port Option |
| 114 | * HP TFT7600 KVM Console US Kit |
| 115 | |
| 116 | == InstaGENI Implementation == |
| 117 | Any configuration or run-time requirements for any of the InstaGENI rack systems will be documented here, when available. |
| 118 | |
| 119 | Each InstaGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, control node, experiment nodes, bare-metal nodes, etc.. See the '''[wiki:GENIRacks/InstaGENICompSetup InstaGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#InstaGENIAdministration InstaGENIRacks Administration]''' tasks. |
| 120 | |
| 121 | == InstaGENI Monitoring == |
| 122 | InstaGENI Racks meet the following [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. InstaGENI Monitoring data is currently being defined and will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP home]. |
| 123 | |
| 124 | == InstaGENI Site Requirements == |
| 125 | InstaGENI racks have the following site requirements: |
| 126 | * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc. |
| 127 | * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components. |
| 128 | * Power requirements - based on site requirements |
| 129 | * Administrative accounts - Setup of site administrator account that will be created on the management/head node. |
| 130 | * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place. Also covers any physical restrictions for the rack delivery. |
| 131 | |
| 132 | If you are interested in becoming an InstaGENI deployment site, please contact us at [mailto:rick.mcgeer@hp.com]. |
| 133 | |
| 134 | ------ |
75 | | = InstaGENI Racks = |
76 | | The InstaGENI project is an effort to deploy low-end, expandable GENI Racks at large number of campuses and is typically found outside the site firewall, more details are to be added as defined. An overview of the project is available from the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/McGeer%20InstaGENI.pdf?format=raw GEC12 InstaGENI Racks] presentation. An [http://groups.geni.net/geni/attachment/wiki/GENIRacksHome/InstaGENI%20Design%20Document%20Rick%202-15%282%29.doc?format=raw InstaGENI Design] document is available. |
77 | | |
78 | | == InstaGENI Components == |
79 | | |
80 | | Each InstaGENI racks delivers a small ProtoGENI cluster with !OpenFlow networking and FOAM aggregate management. This rack includes the following types of systems: |
81 | | * Control Node - Xen Server that runs 3 VMS to provide: |
82 | | * PG boss node, web server, monitoring, and GENI API Server |
83 | | * Local File Server node |
84 | | * FOAM Controller |
85 | | * Experiment Nodes - Five nodes managed by ProtoGENI software stack, which provides boot services, account creation, experimental management, etc. |
86 | | * !OpenFlow Switch - Provides internal routing and data plane connectivity to the GENI backbone (layer 2 and layer 3) |
87 | | * Management Switch - Provides control plane connectivity to the Internet (layer 3) |
88 | | |
89 | | Following are the network connections for an InstaGENI rack: |
90 | | |
91 | | [[Image(GENI-InstaGENI-components.jpg)]] |
92 | | |
93 | | == InstaGENI Specifications == |
94 | | The current hardware components specification for the InstaGENI Racks includes: |
95 | | |
96 | | ''__Compute Resource Specifications__'' 5 HP !ProLiant DL360 G7 Server series hosts to provide the VM Server, Monitoring, Storage, and Application functions: |
97 | | * Control Node - 1 HP !ProLiant DL360 G7 Server, quad-core, single-socket, 12 GB Ram, 4 TB Disk (RAID), and dual NIC |
98 | | * Experiment Nodes - 5 HP !ProLiant DL360 G7 Server, six-core, dual-socket,48GB Ram, 1TB Disk, and dual NIC |
99 | | * Bare Metal Node |
100 | | |
101 | | ''__Network Components__'' X network components to provide access to GENI Backbone and Commodity Internet: |
102 | | * HP !ProCurve 2610 switch, 24 10/100 Mb/s ports, 2 1 Gb/s ports |
103 | | * HP !ProCurve 6600 switch 48 1 Gb/s ports, 4 10 Gb/s ports |
104 | | |
105 | | ''__Misc. Components_'' General purpose hardware also included: |
106 | | * 1 or more HP Intelignt Mod PDU |
107 | | * HP Dedicated iLO Management Port Option |
108 | | * HP TFT7600 KVM Console US Kit |
109 | | |
110 | | == InstaGENI Implementation == |
111 | | Any configuration or run-time requirements for any of the InstaGENI rack systems will be documented here, when available. |
112 | | |
113 | | Each InstaGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, control node, experiment nodes, bare-metal nodes, etc.. See the '''[wiki:GENIRacks/InstaGENICompSetup InstaGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#InstaGENIAdministration InstaGENIRacks Administration]''' tasks. |
114 | | |
115 | | == InstaGENI Monitoring == |
116 | | InstaGENI Racks meet the following [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. InstaGENI Monitoring data is currently being defined and will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP home]. |
117 | | |
118 | | == InstaGENI Site Requirements == |
119 | | InstaGENI racks have the following site requirements: |
120 | | * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc. |
121 | | * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components. |
122 | | * Power requirements - based on site requirements |
123 | | * Administrative accounts - Setup of site administrator account that will be created on the management/head node. |
124 | | * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place. Also covers any physical restrictions for the rack delivery. |
125 | | |
126 | | If you are interested in becoming an InstaGENI deployment site, please contact us at [mailto:rick.mcgeer@hp.com]. |
127 | | |
128 | | -------------------- |
129 | | |
130 | | = ExoGENI Racks = |
131 | | |
132 | | The [http://www.exogeni.net ExoGENI] project is an effort to implement high-performance GENI Racks via a partnership between RENaissance Computing Institute (RENCI), Duke and IBM. ExoGENI racks are assembled and tested by IBM and shipped directly to sites, where they are managed by the RENCI team. ExoGENI racks deliver support for multi-domain cloud structure with flexible virtual networking topologies that allow combining ExoGENI, Meso-scale !OpenFlow and WiMAX resource. An overview of this project was presented at the [http://groups.geni.net/geni/attachment/wiki/GEC12GENIUpdatePlenary/GEC12-ExoGENI-Racks.pptx?format=raw GEC12 ExoGENI Presentation]. Also available are an [https://code.renci.org/gf/download/docmanfileversion/13/691/ExoGENIDesignv1.02.pdf ExoGENI Design] document and an [http://www.cs.duke.edu/~chase/exogeni.pdf ExoGENI white paper]. |
133 | | |
134 | | == ExoGENI Components == |
135 | | |
136 | | An ExoGENI Rack delivers the following types of systems: |
137 | | * Management Switch - An IBM G8052R switched is delivered to allow access to/from the local network provider. |
138 | | * VPN appliance - A Juniper SSG5 provides backup access to manage nodes. |
139 | | * !OpenFlow-enabled switch - An IBM G8264R switch to carry experimental traffic via layer2 to GENI backbone (I2, NLR) and to local OF campus. |
140 | | * Management Node - An IBM x3650 host running Elastic Compute Cloud(EC2) with !OpenStack to provision VMs and running xCat to provision bare-metal nodes. Also runs monitoring functions. |
141 | | * Worker Nodes - Ten IBM 3650 M3 Worker nodes provide both !OpenStack virtualized instances and Bare-metal Xcat nodes |
142 | | * Monitoring Host - None, Monitoring is through Nagios from GMOC. |
143 | | |
144 | | The ExoGENI resources have the following connections: |
145 | | |
146 | | [[Image(GENI-ExoGENI-components.jpg)]] |
147 | | |
148 | | == ExoGENI Specifications == |
149 | | |
150 | | An initial inventory of the ExoGENI Rack hardware components is found [https://docs.google.com/document/d/1hzleT6TNmiDb0YkkgqjXxPFJ37P4O6qApLmXgKJHBZQ/edit?hl=en_US here], which is superseded by the following: |
151 | | |
152 | | ''__Compute Resource__'' A total of 12 hosts are in the rack to provide the Resources, Monitoring, Storage and Application functions: |
153 | | * Management node: 1 IBM x3650 M3, 2x146GB 10K SAS hard drives, 12G RAM, dual-socket 4-core Intel X5650 2.66Ghz CPU, Quad-port 1Gbps adapter |
154 | | * Worker/Bare-Metal nodes: 10 IBM x3650 M3, 1x146GB 10K SAS hard drive +1x500+GB secondary drive, 48G RAM, dual-socket 6-core Intel X5650 2.66Ghz CPU, dual 1Gbps adapter, 10G dual-port Chelseo adapter |
155 | | * Sliverable Storage: 1 IBM DS3512 storage NAS 6x1TB 7200RPM drives |
156 | | |
157 | | ''__Network Components__'' |
158 | | * Management Switch: IBM BNT G8052R 1G client/10G uplink ports - Access to commodity internet. |
159 | | * !OpenFlow Switch: IBM BNT G8264R 10G client/40G uplink ports - Access to GENI backbone. |
160 | | * VPN Appliance: Juniper SSG5 - Backup management access. |
161 | | |
162 | | ''__Misc. Components__'' General purpose hardware included: |
163 | | * IBM PDU based on site power requirements, (GPO=IBM 5897 PDU; RENCI=DPI 5900 PDU) |
164 | | * No UPS included |
165 | | * IBM Local 2x16 Console Manager (LCM16) |
166 | | * IBM 1U 17-inch Flat Panel Console Kit (PN 172317X) |
167 | | |
168 | | == ExoGENI Implementation == |
169 | | Any configuration or run-time requirements for the ExoGENI rack systems will be documented here, when available. |
170 | | |
171 | | Each ExoGENI Rack systems requires a specific setup, which is to captured for each of the rack components: !OpenFlow Switch, Management node, Worker Nodes, etc.. See the '''[wiki:GENIRacks/ExoGENICompSetup ExoGENI Component Setup]''' for details about the required configuration. For help about the configuration settings, please contact [mailto:help@geni.net help@geni.net]. Some examples are available to get insight into '''[wiki:GENIRacksAdministration#ExoGENIAdministration ExoGENI Racks Administration]''' tasks. |
172 | | |
173 | | == ExoGENI Monitoring == |
174 | | Monitoring data for the ExoGENI rack is collected on the management node by a Nagios aggregator and then forwarded to the GMOC. The type of data that will be available is currently being defined. ExoGENI Racks meet the GENI [wiki:GeniRacks#D.MonitoringRequirements monitoring requirements]. ExoGENI Monitoring data will be available at the [https://gmoc-db.grnoc.iu.edu/api-demo GMOC SNAPP Home]. |
175 | | |
176 | | == ExoGENI Site Requirements == |
177 | | ExoGENI racks have the following site requirements: |
178 | | * Network Setup - Define how the rack will connect to the Internet and to the GENI backbones. Ex Regional connections, connection speed, VLANs to be used, etc. |
179 | | * Site Security Requirements- Determine changes required for rack connectivity, such as FV rules, IP filters, etc. |
180 | | * Address assignment for rack components - Define which address, subnet mask, routing is to be configured for the rack components. |
181 | | * Power requirements - based on site requirements |
182 | | * Administrative accounts - Setup of site administrator account that will be created on the management/head node. |
183 | | * Delivery logistics - Details for ''where'' the rack is to be delivered, ''who'' will accept the delivers, and ''when'' the delivery will take place. Also covers any physical restrictions for the rack delivery. |
184 | | |
185 | | If you are interested in becoming an ExoGENI deployment site, please contact us at [mailto:ibaldin@renci.org]. |
186 | | ----- |
187 | | |