ticket,summary,component,version,milestone,type,owner,status,created,_changetime,_description,_reporter
63,incorrect or wrong component_manager_id results in unreported bad state sliver,AM,SPIRAL5,,defect,somebody,new,2012-12-14T10:43:51-05:00,2012-12-16T09:47:50-05:00,"Created a sliver with 2 VMs using the ""routable_control_ip"" tag, the sliver is ready, but the manifest contains no login information.
Here is the sliver manifest:
{{{
}}}
Also rspeclint on the manifest generates the following error:
{{{
$ rspeclint lnroute-manifest-rspec-instageni-gpolab-bbn-com-protogeniv2.xml
Line 8: Failed validation with root at element: Schemas validity error : Element '{http://www.geni.net/resources/rspec/3}link': The attribute 'vlantag' is required but missing.
: rspec
}}}
Is this feature supported? Had syntax changed?
Attaching original Request RSpec.
",lnevers@bbn.com
77,AM V3 reports incorrect error when no local resources exist for the aggregate.,AM,SPIRAL5,,defect,somebody,new,2013-01-07T12:33:10-05:00,2013-03-27T09:44:31-04:00,"In a scenario where the AM API V3 aggregate is queried for a slice which has no resources local to the aggregate, the incorrect error is returned.
Here is the sequence that show the error:
1. Create a slice: ($omni.py createslice lnv3)
2. Describe the slice, which has no resources at the V3 aggregate:
{{{
$ omni.py describe lnv3 -a ig-gpo
INFO:omni:Loading config file /home/lnevers/.gcf/omni_config
INFO:omni:Using control framework pg
INFO:omni:Substituting AM nickname ig-gpo with URL http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/2.0, URN unspecified_AM_URN
ERROR:omni:Describe is only available in AM API v3+. Use ListResources with AM API v2, or specify -V3 to use AM API v3.
lnevers@arendia:~/gcf-2.2-rc3$ omni.py describe lnv3 -a ig-gpo3 -V3
INFO:omni:Loading config file /home/lnevers/.gcf/omni_config
INFO:omni:Using control framework pg
INFO:omni:Substituting AM nickname ig-gpo3 with URL http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/3.0, URN unspecified_AM_URN
INFO:omni:Slice urn:publicid:IDN+pgeni.gpolab.bbn.com+slice+lnv3 expires on 2013-01-08 18:26:11 UTC
INFO:omni:Substituting AM nickname ig-gpo3 with URL http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/3.0, URN unspecified_AM_URN
INFO:omni:Describe Slice urn:publicid:IDN+pgeni.gpolab.bbn.com+slice+lnv3:
WARNING:omni:Got no resource listing from AM http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/3.0
INFO:omni: ------------------------------------------------------------
INFO:omni: Completed describe:
Options as run:
aggregate: ['ig-gpo3']
api_version: 3
framework: pg
Args: describe lnv3
Result Summary: Slice urn:publicid:IDN+pgeni.gpolab.bbn.com+slice+lnv3 expires on 2013-01-08 18:26:11 UTC
Failed to Describe slice urn:publicid:IDN+pgeni.gpolab.bbn.com+slice+lnv3 at AM http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/3.0: Error from Aggregate: code 7. protogeni AM code: 7: Slice credential not provided (PG error log: urn:publicid:IDN+instageni.gpolab.bbn.com+log+13209).
Found description of slivers on 0 of 1 possible aggregates.
INFO:omni: ============================================================
}}}
The error is for slice credential, which is not the proper reason for this failure. Error needs to reflect that the aggregate has no resources for the requested slice.
",lnevers@bbn.com
78,AM V3 reports incorrect error for bad credential type (type geni_sfaBROKEN),AM,SPIRAL5,,defect,somebody,new,2013-01-07T12:43:03-05:00,2013-03-27T10:12:45-04:00,"In a scenario where bad credentials are used (type geni_sfaBROKEN), the aggregate returns the following error:
{{{
No resources from AM http://instageni.gpolab.bbn.com/protogeni/xmlrpc/am/3.0: (Error
from Aggregate: code 1. protogeni AM code: 1: Malformed arguments: Wrong number of
credentials (PG error log: urn:publicid:IDN+instageni.gpolab.bbn.com+log+13056).)
}}}
The message seem to imply that the wrong number of credentials were given which is not the case.
",lnevers@bbn.com
95,Inconsisent naming used for InstaGENI aggregates,AM,SPIRAL5,,defect,somebody,new,2013-02-19T10:15:06-05:00,2013-08-09T10:21:36-04:00,"The current deployment of InstaGENI sites includes the following FOAM and Compute aggregate names:
{{{
Utah: https://foam.utah.geniracks.net and https://utah.geniracks.net
GPO: https://foam.instageni.gpolab.bbn.com and https://instageni.gpolab.bbn.com
Northwestern:https://foam.instageni.northwestern.edu and https://instageni.northwestern.edu
Kentucky: https://foam.lan.sdn.uky.edu:3626/foam/gapi/1 and http://lan.sdn.uky.edu/protogeni/xmlrpc/am/2.0
}}}
There does not seem to be a convention for naming the sites. In the page
http://www.protogeni.net/wiki/RackSetupQuestions there is a suggestion
'' ""Domain for your rack. Something like instageni.foo.bar.edu."" ''
Should we give a stronger indication than the current instructions, so that we may more strongly encourage the use of this naming convention?
",lnevers@bbn.com
115,SSH Keys missing after Joining an Existing Group in Web UI,Administration,SPIRAL5,,defect,somebody,new,2013-05-24T15:55:32-04:00,2013-05-24T15:55:32-04:00,"
When ssh keys are submitted in the ""Join an Existing Project"" web page when requesting a new account they are not available when the account is approved.
Here is the sequence:
1. Request a new account by ""Joining an Existing Group. In the form, I upload ssh keys and select emulab-ops.
2. Once the account is approved, I find that the keys are not in the SSH key page, should they be?
",lnevers@bbn.com
116,Xen host default image is not the same as in Advertisment RSpec,Experiment,SPIRAL5,,defect,somebody,new,2013-08-09T13:42:54-04:00,2013-08-09T13:42:54-04:00,"Ran a test which requested 2 Xen VM without specifying a disk_image to verify the default OS allocation. On the allocated nodes I found:
{{{
xen1:~% cat /etc/issue
Ubuntu 12.04 LTS \n \l
}}}
In the Advertisement RSpec for the Xen server pc3, I see the following default disk_image:
{{{
}}}
The advertisement should match the version actually used by default.
",lnevers@bbn.com
117,Unable to create custom image for FreeBSD82,Experiment,SPIRAL5,,defect,somebody,new,2013-08-09T16:06:00-04:00,2013-08-21T17:16:39-04:00,"
Create custom image failure for FreeBSD82 image. The createimage command details:
{{{
""url"": ""https://boss.utah.geniracks.net/image_metadata.php?uuid=976773ef-012b-11e3-970d-000000000000"",
""urn"": ""urn:publicid:IDN+utah.geniracks.net+image+ch-geni-net:lncustbsd82""
}}}
The email error notification:
{{{
Subject: UTAHGENIRACK: Failed to clone image
Date: Fri, 9 Aug 2013 13:42:52 -0600 (MDT)
From: lnevers@bbn.com
To: lnevers@bbn.com
About to: '/usr/testbed/bin/sshtb -n -o ConnectTimeout=10 -host pcvm3-1 /usr/local/etc/emulab/reboot_prepare'
shutdown: [pid 19642]
Shutdown NOW!
Shutdown NOW!
System shutdown time has arrived
About to: '/usr/testbed/bin/sshtb -n -host pc3 /usr/local/bin/create-image -S 155.98.34.4 -F ch-geni-net/lncustbsd82 -s 1 pcvm3-1 -' as uid 0
pc3: started image capture, waiting up to 72 minutes
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...Container is still running. Waiting ...Container is still running. Waiting ...
Container is still running. Waiting ...
pc3: still waiting ... it has been 2 minutes. Current image size: 0 bytes.
Container is still running. Waiting ...
Container is still running. Waiting ...
Container would not stop!
FAILED: Returned error code 255 generating image ...
}}}
According to Leigh Stroller:
""We need to fix this image or just move to FBSD90-STD.
",lnevers@bbn.com
118,There is no check for lower limit of xen host guest memory size,AM,SPIRAL5,,defect,somebody,new,2013-08-12T11:36:58-04:00,2013-08-12T11:36:58-04:00,"Able to submit a request a for a guest host with memory size of 64 KB.
The sliver did eventually fail, but the request was accepted.
There should be a check to fail for guest memory requests outsize the supported limit.
Addionally, if possible, the error should indicate the valid range of supported guest memory size.
",lnevers@bbn.com
119,Sliver creation for 20 sliver containing 2 VM plus one link each fails. IG-XN-3 Limits tests,Experiment,SPIRAL5,,defect,somebody,new,2013-08-12T15:42:41-04:00,2013-08-12T15:42:41-04:00,"Writing ticket to track resolution of this issue which was initially discussed in email.
Started 20 slivers, 90 seconds apart. Each sliver has 2 Xen VMs slivers with one link.
Slivers took 20-25 minutes for both nodes to be ready (or fail).
Out of the 20 slivers (40 hosts) requested, 16 slivers (25 hosts) failed on creation:
{{{
Nodes: [Node: pcvm3-21] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice3 failed.
Nodes: [Node: pcvm3-24] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice4 failed.
Nodes: [Node: pcvm3-28] [Node: pcvm3-27] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice6 failed.
Nodes: [Node: pcvm3-30] [Node: pcvm3-29] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice7 failed.
Nodes: [Node: pcvm3-31] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice8 failed.
Nodes: [Node: pcvm3-33] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice9 failed.
Nodes: [Node: pcvm3-36] [Node: pcvm3-35] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice10 failed.
Nodes: [Node: pcvm3-38] [Node: pcvm3-37] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice11 failed.
Nodes: [Node: pcvm3-39] [Node: pcvm3-40] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice12 failed.
Nodes: [Node: pcvm3-41] [Node: pcvm3-42] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice13 failed.
Nodes: [Node: pcvm3-43] [Node: pcvm3-44] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice14 failed.
Nodes: [Node: pcvm3-46] [Node: pcvm3-45] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice15 failed.
Nodes: [Node: pcvm3-50] [Node: pcvm3-49] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice17 failed.
Nodes: [Node: pcvm3-52] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice18 failed.
Nodes: [Node: pcvm3-53] [Node: pcvm3-54] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice19 failed.
Nodes: [Node: pcvm3-56] urn:publicid:IDN+ch.geni.net:ln-prj+slice+2vmslice20 failed.
}}}
On 8/12/13 11:01 AM, Leigh Stoller wrote:
> Hi. Yep, see the problem. A write lock on the create path that should be a
> read lock instead. Let me work on this and get back to you.
",lnevers@bbn.com
120,Setup shared node support does not include XEN instructions,Administration,SPIRAL5,,defect,somebody,new,2013-08-12T15:52:24-04:00,2013-08-12T15:52:24-04:00,"As part of the administrative test case, there is a check to redefine the shared node pool.
The instructions at https://wiki.emulab.net/Emulab/wiki/SetupSharedNodes do not include any instructions for how to add/delete a Xen server in the shared pool.
Updated are needed.
",lnevers@bbn.com
123,xen request and allocation mismatch,Experiment,SPIRAL5,,defect,somebody,new,2013-08-12T22:52:33-04:00,2013-08-14T08:37:02-04:00,"In a scenario where one exclusive and one non-exclusive emulab-xen VMs are requested:
{{{
}}}
Both VMs allocated are exclusive according to the manifest:
{{{
....
...
}}}",lnevers@bbn.com
125,Unable to login to Utah DDC procurve2,Administration,SPIRAL5,,defect,somebody,new,2013-08-16T16:25:39-04:00,2013-08-19T09:51:38-04:00,"Writing ticket to track to resolution:
Trying to ""ssh manager@procurve2"" on the Utah DDC boss host, but I am finding that the password in /usr/testbed/etc/switch.pswd does not seem to work.
Is the procurve2 account for this rack using a different password?
The connection failure while using the password in /usr/testbed/etc/switch.pswd:
{{{
[lnevers@boss ~]$ ssh manager@procurve2
We'd like to keep you up to date about:
* Software feature updates
* New product announcements
* Special events
Please register your products now at: www.hp.com/networking/register
manager@procurve2's password:
Permission denied, please try again.
manager@procurve2's password:
Received disconnect from 10.3.1.253: 11: Too many authentication failures.
}}}",lnevers@bbn.com
126,Sliver failures while requesting mulitple VLANs for PG Utah,Experiment,SPIRAL5,,defect,somebody,new,2013-08-16T16:31:57-04:00,2013-08-21T10:22:07-04:00,"Writing ticket to track to resolution:
Not able to exchange traffic over the second link when a stitched two link scenario is run between a GPO Xen host and a Utah Xen host. The same scenario works if the 2 Xen hosts are in the same rack.
Email exchanged about this issue:
{{{
On 8/16/13 9:35 AM, Luisa Nevers wrote:
> Hi Jonathan,
>
> I tried this 2 link topology again this morning and it finally works. I have
> successfully created a stitched sliver between the GPO and Utah racks,
> but only one link works. There is no network connectivity over one of the
> links. The slice is named xen-2link and the hosts allocated are:
>
> - pc5.instageni.gpolab.bbn.com port 30266
> - pc3.utah.geniracks.net port port 30010
>
> Luisa
>
>
> On 8/15/13 6:11 PM, Jonathon Duerig wrote:
>> Luisa, I didn't get a chance to look at this today but will look tomorrow morning when I get in.
>>
>> ---
>> Broad audience or deep message: Pick one.
>>
>>
>> On Thu, 15 Aug 2013, Leigh Stoller wrote:
>>>> Hi Leigh,
>>>
>>> Hi Luisa. I'll defer to Jon on this ...
>>>
>>>> I have been trying to set up stitched sliver from IG GPO to IG Utah with 2 links
>>>> for the past two days, but each attempt always fails with Emulab reporting not
>>>> having a VLAN for one of the links. I am wondering if there is really only one
>>>>VLAN available at Emulab? I just ran the test and here are some of the failure:
>>>> https://www.emulab.net/spewlogfile.php3?logfile=ac1fc8f53ba852a957f8d7a54fd8691f
>>>> https://www.emulab.net/spewlogfile.php3?logfile=49e1c5c2040262428f40bd01dc29ab10
>>>> https://www.emulab.net/spewlogfile.php3?logfile=ac1fc8f53ba852a957f8d7a54fd8691f
>>>>
>>>> Luisa
>>>>
>>>
>>>
}}}
",lnevers@bbn.com
129,AM API v3 performoperationalaction return is 0,AM,SPIRAL5,,defect,duerig@flux.utah.edu,new,2013-09-19T08:57:07-04:00,2013-09-19T08:57:07-04:00,"The return value from AMAPIv3's performoperationalaction call is always '0', not the struct defined by the spec.",ahelsing@bbn.com
13,Supported images availability and documentation,AM,SPIRAL4,,task,ricci@cs.utah.edu,new,2012-05-14T13:04:45-04:00,2013-01-24T14:38:21-05:00,"
On 2/26/12 10:55 PM, Robert Ricci wrote:
>> G.10. Document the default OS Version(s) supported and any known OS restrictions in an InstaGENI rack for bare-metal and VM nodes.
>
> Okay, we will document this. My current plan is to provide default OS images for Ubuntu, CentOS, and FreeBSD. If you have any feedback from experimenters on which Linux distributions they would prefer, that would be helpful.
This ticket tracks the availability to experimenters of the above details:
- The document, which captures supported OS and restrictions.
- The ""default"" images mentioned, initially included
Note: According to listresources there is only one OS available, the ""Standard 32-bit Fedora 15 image"" (urn:publicid:IDN+utah.geniracks.net+image+emulab-ops//FEDORA15-STD).
",lnevers@bbn.com
75,Create Emergency Stop documentation for InstaGENI aggregates,Administration,SPIRAL5,,task,ricci@cs.utah.edu,new,2013-01-03T15:09:20-05:00,2013-02-15T17:57:45-05:00,"Stealing the ticket description from a previous email...
We will need Emergency Stop documentation as part of the [http://groups.geni.net/geni/wiki/GENIRacksHome/AcceptanceTests/InstageniAcceptanceTestsPlan#IG-ADM-7:DocumentationReviewTest IG-ADM-7] acceptance test. The specific wording for the deliverable is: ""Procedure for performing a shutdown operation on any type of sliver on a rack, in support of an Emergency Stop request."" Completing this documentation task is also a prerequisite for performing [http://groups.geni.net/geni/wiki/GENIRacksHome/AcceptanceTests/InstageniAcceptanceTestsPlan#IG-ADM-4:EmergencyStopTest IG-ADM-4].
We don't have any sort of deadline for the document, but ideally the procedure would have been documented and walked through before we deliver any racks to non-developer sites.
The kinds of things we are looking for in the document are as follows:
* How to take a GENI URN as input and map it to resources on the rack
that can be shut down
* Slice URN
* User URN
* The GMOC is also willing to pass you data in other forms
if they have the information that you want available
* The list of hammers (small-to-large) that a site administrator has
when performing an Emergency Stop
* How to get the rack back to a known state with minimal disruption
after an Emergency Stop
If you are feeling ambitious, please add any other notes that you deem useful for a shutdown. At this point, it is OK to list mostly big hammers and iteratively update this document as time permits. Note that the [http://gmoc.grnoc.iu.edu/gmoc/index/documents/geni-proposals/gmoc-noc-support-spiral-4-emergency-stop-workflow.html GMOC's emergency stop workflow] for this spiral is available on their website, and you can use that as a reference for the procedure that you are defining.",tupty@bbn.com
93,Create document capturing outage notification process for site rack maintanance,Administration,SPIRAL5,,task,somebody,new,2013-02-13T14:39:44-05:00,2013-02-13T14:39:44-05:00,"A document is needed to capture the notification process for InstaGENI rack outages. Document should capture the notification process, notification targets (including site contacts). Also if appropriate, timing requirements should also be captured.
Current InstaGENI documentation is being written at http://www.protogeni.net/ProtoGeni/wiki, so it is assumed that this outage notification document will also be at the same location.
",lnevers@bbn.com
100,Resource mapper bug when clientId is the same as the name of a physical host,AM,SPIRAL5,,task,somebody,new,2013-02-21T14:20:06-05:00,2013-02-21T16:16:35-05:00,"Hi I created a slice (transgeo) with two raw nodes and two vms. The topology should look like:
pc1-pc2-vm1-vm2
Notice that pc1 and pc2 are my clientIds and not the actual pc names (I know very unfortunate naming)
pc1 was mapped to pc3.instageni.gpolab.bbn.com and pc2 was mapped to pc4.instageni.gpolab.bbn.com
I asked for a custom Ubuntu image to be loaded in both machines.
pc4.instageni.gpolab.bbn.com has indeed an Ubuntu image loaded while pc43.instageni.gpolab.bbn.com has a fedora one.
The rspec is attached.
The problem is probably caused by the unfortunate naming.
",nriga@bbn.com
103,InstaGENI Software Components Identification Documentation - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:35:10-04:00,2013-03-14T14:35:10-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Software Components Identification Documentation"". Overall goal of the documentation review is to find documents or procedures that:
1. Identify the InstaGENI software components
2. Identify the InstaGENI software components versions
3. Describe system file configuration settings required by the identified components
4. Provide instructions on how to get information about recent changes to the rack software and its configuration. ",lnevers@bbn.com
104,InstaGENI Software Components Functional Description and Access to Source code - Documentation,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:39:18-04:00,2013-03-14T14:39:18-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Software Components Functional Description and Access to Source code"". Overall goal of the documentation review is to find documents or procedures that provide:
1. A description of the InstaGENI software running on a standard rack,
2. Explanation of how to get access to the source code for each component of the standard GENI software.
",lnevers@bbn.com
105,InstaGENI Rack Resources Usage and Policy Documentation - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:41:12-04:00,2013-03-14T14:41:12-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Rack Resources Usage and Policy Documentation"".
No documentation was found for rack resource usage policy. ",lnevers@bbn.com
106,InstaGENI Rack Resources Administration - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:44:59-04:00,2013-03-14T14:44:59-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Rack Resources Administration"".
The goal of this documentation evaluation is to find documentation that:
1. Provides a description of the expected state of all the GENI experimental resources in the rack,
2. Documents how to determine the state of an experimental resource
3. Documents what state is expected for an unallocated bare metal node.
No information was found about InstaGENI rack resource administration. ",lnevers@bbn.com
107,InstaGENI Rack Administrative procedure for rack operations - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:47:34-04:00,2013-03-14T14:47:34-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Rack Administrative procedure for rack operations"". Goal for documentation review is to find:
- Rack operator procedures
- Instructions for changing IP addresses on all rack components
No information was found for operations.",lnevers@bbn.com
108,InstaGENI Rack Administrative procedure for health status - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:49:50-04:00,2013-03-14T14:49:50-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Rack Administrative procedure for health status"". The overall goal of the documentation review is to find documentation that:
1. provides a procedure for performing comprehensive health checks for a rack
2. (or, if those health checks are being run automatically, how to view the current/recent results).
No such procedure or status viewing instructions were found.",lnevers@bbn.com
109,Site Specific InstaGENI Deployment Data - Documentation requirements,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T14:54:48-04:00,2013-03-14T14:54:48-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""Site Specific InstaGENI Deployment Data"".
The goal of this evaluation is to find documentation that:
1. Provides per-site deployment details/documentation.
2. Documentation should be prepared before sites are installed and kept updated after installation to reflect any changes or upgrades after delivery.
3. Text, network diagrams, wiring diagrams and labeled photos are all acceptable for site documents.
4. Per-site documentation should include the following items for each site:
a. Part numbers and quantities of PDUs, with NEMA input power connector types, and an inventory of which equipment connects to which PDU.
b. Physical network interfaces for each control and data plane port that connects to the site's existing network(s), including type, part numbers, maximum speed etc. (eg. 10-GB-SR fiber)
c. Public IP addresses allocated to the rack, including: number of distinct IP ranges and size of each range, hostname to IP mappings which should be placed in site DNS, whether the last-hop routers for public IP ranges subnets sit within the rack or elsewhere on the site, and what firewall configuration is desired for the control network.
d. Data plane network connectivity and procedures for each rack, including core backbone connectivity and documentation, switch configuration options to set for compatibility with the L2 core, and the site and rack procedures for connecting non-rack-controlled VLANs and resources to the rack data plane. A network diagram is highly recommended.",lnevers@bbn.com
110,InstaGENI Rack Site Operations documentation - Documentation requirement,Experiment,SPIRAL5,,task,somebody,new,2013-03-14T15:00:15-04:00,2013-03-14T15:00:15-04:00,"The test case IG-ADM-7 checks for availability of InstaGENI rack documentation. This ticket captures the requirement to deliver ""InstaGENI Rack Site Operations documentation"".
The goal of this documentation evaluation is to find documentation that provides a Technical plan for handing off primary rack operations to site operators at all sites.
No plan found defining hand off plan for the InstaGENI racks.",lnevers@bbn.com
121,No Monitoring statistics for Xen servers in GMOC monitoring data.,Monitoring,SPIRAL5,,task,somebody,new,2013-08-12T16:00:52-04:00,2013-08-12T16:00:52-04:00,"
As part of the test case IG-XN-12: Xen Resource Monitoring, reviewing of the GMOC aggregate data takes place to find the following for the new Xen resources:
1. Identify Xen VMs resources
2. CPU Statistics
3. Disk Statistics
4. Virtualization statistics
Reviewed data for GPO and Utah InstaGENI Aggregates at the at GMOC and CPU, Disk and Virtualization statistics are not found.
There is information that identifies the server as a VM resource, but there is no information about the type of VM resource (openvz vs xen) in use.
",lnevers@bbn.com
130,AM API v3 geni_restart isn't restarting nodes?,AM,SPIRAL5,,task,duerig@flux.utah.edu,new,2013-09-19T08:58:07-04:00,2013-09-19T08:58:07-04:00,"Leigh reported that issuing a geni_restart did not result in the expected log messages indicating that the node actually restarted.
Is this working?",ahelsing@bbn.com
131,ExoSM resources do not appear in Flack,Experiment,SPIRAL5,,task,duerig@flux.utah.edu,new,2013-09-19T09:00:19-04:00,2013-09-19T09:00:19-04:00,Resources belonging to the ExoSM do not appear in Flack.,ahelsing@bbn.com
134,"AM API V3 provision fails with ""this is not your ticket error""",Experiment,SPIRAL6,,,somebody,new,2014-05-22T07:18:57-04:00,2014-05-22T07:18:57-04:00,"The AM API V3 provision call no longer works:
{{{
$ omni.py provision -a am-undertest lnv3 ./request.xml -V3
07:14:40 INFO omni: Loading agg_nick_cache file '/home/lnevers/.gcf/agg_nick_cache'
07:14:40 INFO omni: Loading config file omni_config
07:14:40 INFO omni: Using control framework portal
07:14:40 INFO omni: Member Authority is https://ch.geni.net/MA (from config)
07:14:40 INFO omni: Slice Authority is https://ch.geni.net/SA (from config)
07:14:40 INFO omni: Substituting AM nickname am-undertest with URL https://boss.instageni.gpolab.bbn.com:12369/protogeni/xmlrpc/am/3.0, URN unspecified_AM_URN
07:14:40 INFO omni: Slice urn:publicid:IDN+ch.geni.net:ln-test+slice+lnv3 expires on 2014-05-29 11:13:21 UTC
07:14:40 INFO omni: Provision slivers in slice urn:publicid:IDN+ch.geni.net:ln-test+slice+lnv3 at gpo3-ig
07:14:42 WARNING omni: Provision of slivers in slice urn:publicid:IDN+ch.geni.net:ln-test+slice+lnv3 at gpo3-ig failed: Error from Aggregate: code 3. protogeni AM code: 3: This is not your ticket (PG log url - look here for details on any failures: https://boss.instageni.gpolab.bbn.com/spewlogfile.php3?logfile=69c9562fa8d6e54a32f8a1113bb4a0de).
07:14:42 INFO omni: ------------------------------------------------------
07:14:42 INFO omni: Completed provision:
Args: provision lnv3 ./request.xml
Result Summary: Provision of slivers in slice urn:publicid:IDN+ch.geni.net:ln-test+slice+lnv3 at gpo3-ig failed: Error from Aggregate: code 3. protogeni AM code: 3: This is not your ticket (PG log url - look here for details on any failures: https://boss.instageni.gpolab.bbn.com/spewlogfile.php3?logfile=69c9562fa8d6e54a32f8a1113bb4a0de).
Provision slivers in slice urn:publicid:IDN+ch.geni.net:ln-test+slice+lnv3 failed at https://boss.instageni.gpolab.bbn.com:12369/protogeni/xmlrpc/am/3.0
07:14:42 INFO omni: ======================================================
}}}
This problem is repeatable on every attempt.",lnevers@bbn.com
136,AM API V3 Describe reports internal error for slice that does not exist at aggregate,AM,SPIRAL6,,,somebody,new,2014-05-22T08:40:50-04:00,2014-05-22T08:42:22-04:00,"The AM API V3 describe command causes an internal error when executed at an aggregate that does not have any resources for the slice:
{{{
$ omni.py describe nonexist -V3 -a nysernet-ig
08:37:37 INFO omni: Loading agg_nick_cache file '/home/lnevers/.gcf/agg_nick_cache'
08:37:37 INFO omni: Loading config file omni_config
08:37:37 INFO omni: Using control framework portal
08:37:37 INFO omni: Member Authority is https://ch.geni.net/MA (from config)
08:37:37 INFO omni: Slice Authority is https://ch.geni.net/SA (from config)
08:37:37 INFO omni: Substituting AM nickname nysernet-ig with URL https://instageni.nysernet.org:12369/protogeni/xmlrpc/am/2.0, URN urn:publicid:IDN+instageni.nysernet.org+authority+cm
08:37:37 INFO omni: Slice urn:publicid:IDN+ch.geni.net:ln-test+slice+nonexist expires on 2014-05-29 12:37:26 UTC
08:37:37 INFO omni: Describe Slice urn:publicid:IDN+ch.geni.net:ln-test+slice+nonexist:
08:37:37 WARNING omni: Requested API version 3, but AM https://instageni.nysernet.org:12369/protogeni/xmlrpc/am/2.0 uses version 2. Same aggregate talks API v3 at a different URL: https://boss.instageni.nysernet.org:12369/protogeni/xmlrpc/am/3.0
08:37:38 INFO omni: Switching AM URL to match requested version
08:37:40 ERROR omni: Call for Describe slice urn:publicid:IDN+ch.geni.net:ln-test+slice+nonexist at https://instageni.nysernet.org:12369/protogeni/xmlrpc/am/2.0 failed. Server says:
08:37:40 WARNING omni: Got no resource listing from AM nysernet-ig3
08:37:40 INFO omni: ------------------------------------------------------
08:37:40 INFO omni: Completed describe:
Args: describe nonexist
Result Summary: Slice urn:publicid:IDN+ch.geni.net:ln-test+slice+nonexist expires on 2014-05-29 12:37:26 UTC
Failed to Describe slice urn:publicid:IDN+ch.geni.net:ln-test+slice+nonexist at AM nysernet-ig3: (missing result)
Found description of slivers on 0 of 1 possible aggregates.
08:37:40 INFO omni: ======================================================
}}}",lnevers@bbn.com
137,vlanRangeAvailability ignored when request 'any',AM,SPIRAL6,,,somebody,new,2014-05-30T15:18:51-04:00,2014-05-30T15:18:51-04:00,"When we request vlan tag 'any', the vlanRangeAvailability value is ignored - and the AM can (and does sometimes at least) pick a tag that is specifically excluded in the vlanRangeAvailability set.
",ahelsing@bbn.com
138,Confirmation test sites missing compute aggregate monitoring,Administration,SPIRAL6,,,asydney@bbn.com,assigned,2014-06-02T08:56:39-04:00,2014-10-03T13:39:07-04:00,"Running New Site Confirmation Tests for NPS and checked GMOC for monitored NPS aggregates:
https://gmoc-db.grnoc.iu.edu/protected-openid/index.pl?method=aggregates&search=nps
Found that there is no Compute aggregate for NPS.
There is a FOAM aggregate.",lnevers@bbn.com
23,utah.control.geniracks.net hostname mismatch,Administration,SPIRAL4,IG-ADM-1,task,ricci@cs.utah.edu,assigned,2012-05-16T23:52:58-04:00,2013-01-14T16:59:21-05:00,"In public DNS (testing from BBN internal network), the control host is called `utah.control.geniracks.net`:
{{{
$ host utah.control.geniracks.net
utah.control.geniracks.net has address 155.98.34.2
$ host control.utah.geniracks.net
Host control.utah.geniracks.net not found: 3(NXDOMAIN)
}}}
On the host itself, the hostname is `control.utah.geniracks.net`:
{{{
control,[~],21:48(0)$ hostname
control.utah.geniracks.net
control,[~],21:51(0)$ ifconfig | grep 155
inet addr:155.98.34.2 Bcast:155.98.34.255 Mask:255.255.255.0
}}}
This is likely to cause confusion, and should probably be standardized on either `.control.geniracks.net` or `control..geniracks.net`.
",chaos@bbn.com
50,sliver manifest does not pass rspeclint,AM,SPIRAL4,IG-EXP-2,defect,somebody,new,2012-09-06T12:51:34-04:00,2013-03-27T09:17:56-04:00,"
The rspeclint command shows the following errors for a sliver manifest:
{{{
$ rspeclint lninsta-manifest-rspec-boss-utah-geniracks-net-protogeniv2.xml
Line 8: Failed validation with root at element: Schemas validity error :
Element '{http://www.geni.net/resources/rspec/3}link': The attribute 'vlantag' is required but missing.
Schemas validity error : Element '{http://www.geni.net/resources/rspec/3}link':
The attribute 'vlantag' is required but missing. : rspec
$
}}}
Attaching the manifest file. The sliver is a 2 VMs setup with 2 links between them, also attaching the RSpec used to create the sliver.",lnevers@bbn.com
48,creating a 50 VM sliver causes failure - Fault -32400: 'XMLRPC Server Error',Experiment,SPIRAL4,IG-EXP-3,defect,somebody,new,2012-08-20T16:40:49-04:00,2012-08-20T16:40:49-04:00,"
Created one 50 VM sliver (urn:publicid:IDN+pgeni.gpolab.bbn.com+slice+2exp-50vm) with omni. The sliver creation completed without error, but a subsequent sliver status ( approx 5 minutes after creation) showed the following error:
{{{
Failed to get SliverStatus on 2exp-50vm at AM http://utah.geniracks.net/protogeni/xmlrpc/am:
}}}
AM is still responding to other commands, such as list resources.
",lnevers@bbn.com
62,rspeclint for sliver manifest reports unkown namespace,AM,SPIRAL5,,defect,somebody,new,2012-12-14T07:32:42-05:00,2013-03-27T08:46:53-04:00,"A sliver manifest does not pass rspeclint and reports:
{{{
Ignoring node rs:vnode with unknown namespace http://www.protogeni.net/resources/rspec/ext/emulab/1
Ignoring node rs:vnode with unknown namespace http://www.protogeni.net/resources/rspec/ext/emulab/1
Ignoring node rs:vnode with unknown namespace http://www.protogeni.net/resources/rspec/ext/emulab/1
Ignoring node rs:vnode with unknown namespace http://www.protogeni.net/resources/rspec/ext/emulab/1
}}}
Attaching RSpec
",lnevers@bbn.com
83,Several sliver in GPO InstaGENI have not configured the requested client_id value,Experiment,SPIRAL5,,defect,somebody,new,2013-01-08T10:58:36-05:00,2013-03-27T09:05:23-04:00,"Several sliver requests in the GPO InstaGENI Rack this morning have resulted in nodes which did not have the ''client_id'' requested in the RSpec. As an example, the sliver IG-EXP-6-exp2 in the GPO rack requested the following client_ids:
{{{
}}}
But the nodes that are available are named ""vhost1"" and ""pc5"" and not as requested ""GPO-VM"" and ""GPO-PC"".
This has occurred several times on the GPO rack, but has not happened with requests to the Utah Rack. Is this an issue?
",lnevers@bbn.com
87,Issuing reboot on GPO and Utah IG nodes results in node not available,AM,SPIRAL5,,defect,somebody,reopened,2013-01-15T11:30:15-05:00,2013-01-15T14:14:18-05:00,"
In several scenario, both at Utah and GPO InstaGENI, issuing a restart causes for the node to not come back up:
{{{
[lnevers@top ~]$ sudo shutdown -r now
Failed to talk to init daemon.
[lnevers@top ~]$ Connection to pc2.instageni.gpolab.bbn.com closed by remote host.
Connection to pc2.instageni.gpolab.bbn.com closed.
lnevers@arendia:~/gcf-2.2-rc4$ sleep 360
lnevers@arendia:~/gcf-2.2-rc4$ ssh -p 30780 lnevers@pc2.instageni.gpolab.bbn.com
ssh: connect to host pc2.instageni.gpolab.bbn.com port 30780: Connection refused
}}}
There are currently 4 nodes that are part of the slice IG-EXP-7 and were restarted and 30 minutes later are still not accepting ssh connections:
* pc2.instageni.gpolab.bbn.com port 30780
* pc2.instageni.gpolab.bbn.com port 30779
* pc3.utah.geniracks.net port 32060
* pc3.utah.geniracks.net port 32058
After the restart, the nodes am_status changes from ""ready"" to ""notready"".
",lnevers@bbn.com
135,AM API V3 GetVersion geni_single_allocation is a string '1' and should be an int,AM,SPIRAL6,,,somebody,new,2014-05-22T07:58:29-04:00,2014-05-22T07:58:29-04:00,"The GetVersion return for geni_single_allocation is a string '1' and should be an int. The field is defined as """" (see http://groups.geni.net/geni/wiki/GAPI_AM_API_V3/CommonConcepts#OperationsonIndividualSlivers), which is an int not a string.
",lnevers@bbn.com
14,Advertisment RSpec reports two default images,Experiment,SPIRAL4,,defect,somebody,new,2012-05-14T16:07:46-04:00,2013-03-27T08:39:42-04:00,"
Tracking the default image duplication in the advertisement RSpec separately from the OS support ticket.
On 5/14/12 2:12 PM, Jonathon Duerig wrote:
> I'll take a look at this later today or tomorrow. Odd, but should be harmless.
>
>> On Mon, 14 May 2012, Leigh Stoller wrote:
>>> 2. Is there a reason why FEDORA-15 is reported twice as ""default=true"" ?
>>
>> Not sure; a question for Jon ...
>>
>> Lbs
",lnevers@bbn.com
139,mkadmin.pl inconsistent permissions,Administration,SPIRAL6,,,asydney@bbn.com,accepted,2014-09-18T19:27:23-04:00,2014-09-19T10:43:30-04:00,"The script mkadmin.pl has inconsistent permissions on the control, foam and flowvisor hosts.
On the control it has 755 permissions, while on the foam and flowvisor hosts it has 744 permissions. The permissions should be consistent, preferably 755 so command line completion can match the file.
",lnevers@bbn.com