FAQ

1. CBTOOL fails to start.

  • Check if the hostname and IP address of CBTOOL Orchestrator node have been added in the /etc/hosts. CBTOOL needs to be able to resolve its own IP address into a name (any name will do). An error message indicating name resolution failure will be printed out
  • Check if the default ports used by CBTOOL - 7070 for the API, 8080 for the GUI and 5114 for logging - are not already taken. An error message indicating that was impossible to bind to the any of these ports will be printed out

2. CBTOOL fails to communicate with its own API.

  • By default the attribute MANAGER_IP, on the CBTOOL’s cloud configuration file (always named as ~/osgcloud/cbtool/configs/LINUXUSERNAME_cloud_definitions.txt) is set to $IP_AUTO. When it finds this special value, CBTOOL attempts to auto-discover the IP address on the Orchestrator. However, if your CBTOOL Orchestrator node has more than one network interface card (or even multiple IPs on a single interface), CBTOOL is unable to automatically decide on which one should be used

    In that case, set an actual IP address (selected by the tester among the many found on the Orchestrator node) on the cloud configuration file

    vi ~/osgcloud/cbtool/configs/ubuntu_cloud_definitions.txt
    MANAGER_IP = IPADDRESS_OF_INTERFACE_FOR_ORCHESTRATOR_NODE
    

    The above snippet assumes that the Linux user under which CBTOOL is running is ubuntu. If Linux user has a different name, the configuration file will have a different starting name (s/ubuntu/LINUXUSERNAME)

  • If cloud under test is an OpenStack cloud, the tester must set the host name of cloud controller (nova API) in /etc/hosts file of the benchmark harness machine

3. Attempting to attach a single VM/Container instance through CBTOOL’s CLI - for testing purposes - results in an error message. How can I debug?

  • Restart CBTOOL in debug mode with ~/osgcloud/cbtool/cb --hard_reset -v 10

  • Instruct CBTOOL to deploy the single VMs (or Containers) without attempting to establishing contact to these, by running the following command on the CLI

    vmdev
    
  • After that, just run

    vmattach cassandra
    

    or

    vmattach hadoopmaster
    

    CBTOOL will output each command that would have been executed against the VM/Containers. At this point, you can just execute the commands yourself and check for error messages.

    NOTE: When running the script cb_post_boot.sh directly on the VM/Container, it should complete succesfully and promptly (e.g., within 90 seconds). If the script hangs, the VM/Container is probably having trouble reaching either the Orchestrator node (specifically on TCP ports 6379,27107 and UDP port 5114) or the NTP server.

  • Once done, do not forget to disable the debug mode by issuing the command CLI

    vmundev
    

    NOTE: There is a small utility, called “cbssh”, that could be used to directly connect to the VMs. To try it, just run the following on a bash prompt

    cd ~/cbtool; ~/cbtool/cbssh vm_1
    

    and you should be able to login on the node.

4. Attempting to attach a single Application Instance (AI) through CBTOOL’s CLI - for testing purposes - results in an error message. How can I debug?

  • Restart CBTOOL in debug mode with ~/osgcloud/cbtool/cb –hard_reset -v 10

  • Instruct CBTOOL to deploy a single Application Instance (AI) without running the actual configuration/startup scripts on its multiple VMs/Containers, by running the following command on the CLI

    appdev
    
  • After that, just run

    aiattach cassandra_ycsb
    

    or

    aiattach hadoop
    

    CBTOOL will output each command that would have been executed against each one of the VMs/Containers that compose the VApp. At this point, you can just execute the command yourself and check what could be wrong.

    IMPORTANT:Please note that some commands should be executed in parallel (or at least in a non-sequential manner) and therefore might required multiple prompts for overlapping execution

  • Once done, do not forget to disable the debug mode by issuing the command

    appundev
    

5. Baseline Application Instances are deployed successfully, but fail to produce any application metric samples.

  • Restart CBTOOL in debug mode with ~/osgcloud/cbtool/cb –hard_reset -v 10

  • Instruct CBTOOL to fully deploy the Virtual Application instance, but do not start the actual load generation by running the following command on the CLI

    appnoload
    
  • After that, just run

    aiattach cassandra_ycsb
    

    or

    aiattach hadoop
    

    At the very end, CBTOOL will output a message such as “Load Manager will NOT be automatically started on VM NAME during the deployment of VAPP NAME...”.

  • Obtain a list of VMs with the command

    vmlist
    
  • Login on the instance with the role “yscb” (in case of cassandra_ycsb) or “hadoopmaster” (in case of hadoop) using CBTOOL’s helper utility, e.g.

    cd ~/cbtool; ~/osgcloud/cbtool/cbssh vm_4
    
  • Try to run the “load generation” script directly on the instance

    ~/cbtool/cb_ycsb.sh workloadd 1 1 1
    

    or

    ~/cbtool/cb_hadoop_job.sh kmeans 1 1 1
    
  • If the “load generation” script completes successfully, then try to run the Load Manager daemon in debug mode

    /usr/local/bin/cbloadman
    

    Watch for errors in the Load Manager process, displayed directly on the terminal.

  • Once done, do not forget to disable the debug mode by issuing the command

    appload
    

6. Baseline phase hangs. What is happening?

There are couple of reasons why baseline phase can hang:

  • NTP server is not running, or NTP server is misconfigured in osgcloud_rules.yaml
  • The configured image is unable to run CBTOOL post-boot scripts (check previous question on this faq)
  • CBTOOL is unable to reach instances because instances do not have direct network connectivity with the CBTOOL machine. In this case, setup a jump box, and configure CBTOOL to use jump box.

7. How can I check if CBTOOL scripts are failing?

CBTOOL logs are stored in /var/log/cloudbench . Search for errors as follows:

cd /var/log/cloudbench/
tail -f /var/logs/cloudbench/LINUXUSERNAME_remotescripts.log

Search for the AI or the instance name for which there are errors.

8. Supporting evidence collection is not working for baseline.

Please make sure that the following parameters are set correctly:

instance_user: cbuser
instance_keypath: HOMEDIR/osgcloud/cbtool/credentials/id_rsa

Please also make sure that the permissions of the private key are set correctly, that is:

chmod 400 id_rsa

9. When using CBTOOL with the OpenStack Cloud Adapter, can I instruct it to create each new Application Instance on its own tenant/network/subnet/router?

CBTOOL has the ability to execute generic scripts at specific points during the VM/Container attachment (e.g., before the provision request is issued to the cloud, after the VM is reported as “started” by the cloud). A small example script which creates a new keystone tenant, a new pubkey pair, a new security group and a neutron network, subnet and router was made available under the “scenarios/util” directory.

To use it with Application Instances (i.e., ALL VMs/Containers belonging to an AI on its own tenant/network/subnet/router) issue the following command on the CLI

cldalter ai_defaults execute_script_name=/home/cbuser/cloudbench/scenarios/scripts/openstack_multitenant.sh

After this each new AI that is attached will execute the aforementioned script. You can test it out with the following command

aiattach nullworkload default default none none execute_provision_originated

Please note that the commands listed on the script will be executed from the Orchestrator node, and thus require the existence of all relevant OpenStack CLI clients (e.g., openstack, nova and neutron) present there.

10. What is the difference between maximum_ais and reported_ais in the benchmark’s configuration file osgcloud_rules.yaml ?

This benchmark is designed to stress your cloud’s control plane as much as the data plane. When applications are being submitted to your cloud, during a compliant run, they are required to arrive at the cloud during a random interval between 5 and 10 minutes. If your cloud’s AI empirical provisioning time (including all the VMs participating with that AI / application) is less than or equal to the average arrival times from that interval, then the benchmark will likely always be able to collect data in a timely manner from the vast majority of the AIs sent to your cloud. Basically, VMs won’t get backed up and will get services as fast as the benchmark is throwing them at your cloud. This is (currently) hard to achieve — every VM would likely need to be provisioned well under a minute, and be provisioned in parallel for this to happen – and given that, on average, the two types of AIs that SPEC Cloud uses contain as many as 7 virtual machines in each type of AI. On the other hand, if your cloud’s AI provisioning time (and all participating VMs) are slower than that interval, you will always have an ever-increasing backlog of virtual machines waiting to complete their deployment against your cloud. In this situation, the benchmark will still have a compliant run, but will terminate early without collecting application performance results from all of the AIs. This behavior is normal, expected and compliant. So, don’t be surprised in the later scenario that the computed score doesn’t match the AI limit you set out to achieve. As long is the score is stable and doesn’t fluctuate much (within 10%), the benchmark is doing what it is supposed to be doing.

In either of the aforementioned scenarios, the variables maximum_ais and reported_ais are designed to create consistency in the results computed by SPEC Cloud. Thus, when choosing to prepare a SPEC Cloud submission report, you have a choice between:

Scenario a) Your cloud’s AI provisioning time is faster than the arrival rate (less than or equal to the 5/10 minute interval) ===> The benchmark will likely terminate by reaching reported_ais first.

Scenario b) Your cloud’s AI provisioning time is slower than the arrival rate (greater than the 5/10 minute interval) ===> The benchmark will terminate by reaching maximum_ais first.

In scenario a, you will likely want to set reported_ais to a lower number than maximum_ais, so that SPEC Cloud only calculates a final score based on a consistent number of AIs created by the benchmark. SPEC Cloud does not (and does not pretend to) calculate this number for you —- only the submitter can really choose this number from empirical experience.

In scenario b, you will likely want to set reported_ais equal to maximum_ais so that as many AIs as possible count towards the final score calculated by SPEC Cloud. Public clouds with limits on their API request rates or that throttle the number of simultaneous VMs that can complete at the same time due to DDoS mitigation or load balancing techniques will likely fall into this category.

11. What does daemon_parallelism and attach_parallelism in osgcloud_rules.yaml mean?

Some clouds, in particular public clouds, enforce artificial limits on to protect abuse or denial of service. In such situations, users need the ability to control how fast the benchmark hits the cloud’s API.

====> daemon_parallelism: This controls the maximum simultaneous outstanding number of AIs that are actively being provisioned at the same time (not the number that have been issued.) For a compliant run, the benchmark deploys AIs between 5 and 10 minutes, but if the cloud has customer limits and cannot absorb this load, setting this value can help.

====> attach_parallelism: This controls the maximum simultaneous outstanding number of VMs in a single AI that are actively being provisioned (not the number that the AI is ‘expecting’ to provision). This value can be changed if the cloud needs to be further protected due to abuse or denial of service limits imposed by the cloud, which otherwise cannot be changed for any user.

Lowering the limits set for these parameters may cause CBTOOL to hold issuing request for new application instances if the number of in flight application instances equals the parameter set in daemon_parallelism. If this happens, the interarrival time between application instances may exceed the limit prescribed by the benchmark, that is, 600s. Such a result is considered non-compliant.

12. Verify rules indicate an ‘ERROR’ in rules. Shall I proceed with the experiment?

If a run is started with the existing values in osgcloud_rules.yaml, you may see an error like this:

2016-08-16 15:53:46,270 INFO process ERROR: results->support_evidence->instance_support_evidence is False. Result will be non-compliant.
2016-08-16 15:53:46,270 INFO process ERROR: results->support_evidence->cloud_config_support_evidence is False. Result will be non-compliant.

In SPEC Cloud kit, supporting evidence collection for instances and cloud is set to False. It must be enabled for a compliant run.

If the run is for internal results, there is no need to update the value for these parameters.

13. I am running baseline phase for testing and do not want the instances created during baseline phase to be destroyed.

A compliant run requires that instances created during every iteration of baseline are deleted immediately after the iteration. This is already implemented in the benchmark all_run script.

For testing, to disable instance deletion after each iteration of baseline phase, set the following flag in osgcloud_rules.yaml:

baseline:
  destroy_ai_upon_completion: false

14. How do I check the detailed timeline for deployment of a given Application Instance (AI)?

A very detailed timeline for the deployment of each AI can be found on ~/results/EXPID/perf/EXPIDELASTICITY/<USERNAME>_operations.log

An illustrative example for a 2-VM Application Instance (using the “iperf” workload, not part of SPECCloud) follows:

cbuser@cborch:~/cloudbench$ cat /var/log/cloudbench/cbuser_operations.log | grep ai_1

Aug 16 14:16:03 cborch.public [2016-08-16 14:16:03,865] [DEBUG] active_operations.py/ActiveObjectOperations.pre_attach_ai TEST_cbuser - Starting the attachment of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,302] [DEBUG] active_operations.py/ActiveObjectOperations.pre_attach_vm TEST_cbuser - Starting the attachment of vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,312] [DEBUG] base_operations.py/ActiveObjectOperations.admission_control TEST_cbuser - Reservation for vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) was successfully obtained..

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,326] [DEBUG] active_operations.py/ActiveObjectOperations.pre_attach_vm TEST_cbuser - Starting the attachment of vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,336] [DEBUG] base_operations.py/ActiveObjectOperations.admission_control TEST_cbuser - Reservation for vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) was successfully obtained..

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,744] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_ready TEST_cbuser - Waiting for vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), to start...

Aug 16 14:16:04 cborch.public [2016-08-16 14:16:04,748] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_ready TEST_cbuser - Waiting for vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), to start...

Aug 16 14:16:05 cborch.public [2016-08-16 14:16:05,745] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_ready TEST_cbuser - Check if vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) has started by querying thecloud directly.

Aug 16 14:16:05 cborch.public [2016-08-16 14:16:05,750] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_ready TEST_cbuser - Check if vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) has started by querying thecloud directly.

Aug 16 14:16:05 cborch.public [2016-08-16 14:16:05,756] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_boot TEST_cbuser - Trying to establish network connectivity to vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:05 cborch.public [2016-08-16 14:16:05,757] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_boot TEST_cbuser - Trying to establish network connectivity to vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:06 cborch.public [2016-08-16 14:16:06,758] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_boot TEST_cbuser - Assuming that vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) is booted after waiting for 0 seconds.

Aug 16 14:16:06 cborch.public [2016-08-16 14:16:06,759] [DEBUG] shared_functions.py/PdmCmds.wait_for_instance_boot TEST_cbuser - Assuming that vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) is booted after waiting for 0 seconds.

Aug 16 14:16:06 cborch.public [2016-08-16 14:16:06,887] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Checking ssh accessibility on vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9): ssh -p 10001 fedora@9.2.211.203 "/bin/true"...

Aug 16 14:16:06 cborch.public [2016-08-16 14:16:06,888] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Checking ssh accessibility on vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9): ssh -p 10002 fedora@9.2.211.203 "/bin/true"...

Aug 16 14:16:07 cborch.public [2016-08-16 14:16:07,005] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Checked ssh accessibility on vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)

Aug 16 14:16:07 cborch.public [2016-08-16 14:16:07,006] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Bootstrapping vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9): creating file cb_os_paramaters.txt in "fedora" user's home dir on IP address 9.2.211.203...

Aug 16 14:16:07 cborch.public [2016-08-16 14:16:07,010] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Checked ssh accessibility on vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)

Aug 16 14:16:07 cborch.public [2016-08-16 14:16:07,010] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Bootstrapping vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9): creating file cb_os_paramaters.txt in "fedora" user's home dir on IP address 9.2.211.203...

Aug 16 14:16:09 cborch.public [2016-08-16 14:16:09,375] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Bootstrapped vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)

Aug 16 14:16:09 cborch.public [2016-08-16 14:16:09,375] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Sending a copy of the code tree to vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:09 cborch.public [2016-08-16 14:16:09,945] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Sent a copy of the code tree to vm_1 (F51F167D-7962-51B6-93CA-8E99A5D70748), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:10 cborch.public [2016-08-16 14:16:10,378] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Bootstrapped vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)

Aug 16 14:16:10 cborch.public [2016-08-16 14:16:10,378] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Sending a copy of the code tree to vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:10 cborch.public [2016-08-16 14:16:10,952] [DEBUG] active_operations.py/ActiveObjectOperations.post_attach_vm TEST_cbuser - Sent a copy of the code tree to vm_2 (FD0EA8D8-F89C-5624-8FC3-EF6C6574E367), part of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9), on IP address 9.2.211.203...

Aug 16 14:16:11 cborch.public [2016-08-16 14:16:11,400] [DEBUG] base_operations.py/ActiveObjectOperations.admission_control TEST_cbuser - Reservation for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) was successfully obtained..

Aug 16 14:16:11 cborch.public [2016-08-16 14:16:11,464] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - Performing generic application instance post_boot configuration on all VMs belonging to ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:16 cborch.public [2016-08-16 14:16:16,985] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - The generic post-boot "setup" scripts for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) completed with status 0 after 5 seconds

Aug 16 14:16:16 cborch.public [2016-08-16 14:16:16,993] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - Running application-specific "setup" configuration on all VMs belonging to ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:16 cborch.public [2016-08-16 14:16:16,996] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - QEMU Scraper will NOT be automatically started during the deployment of ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9)...

Aug 16 14:16:21 cborch.public [2016-08-16 14:16:21,502] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - The application-specific "setup" scripts for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) completed with status 0 after 5/5 seconds

Aug 16 14:16:22 cborch.public [2016-08-16 14:16:22,016] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - The application-specific "setup" scripts for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) completed with status 0 after 1/6 seconds

Aug 16 14:16:24 cborch.public [2016-08-16 14:16:24,032] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - The application-specific "setup" scripts for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) completed with status 0 after 2/8 seconds

Aug 16 14:16:24 cborch.public [2016-08-16 14:16:24,045] [DEBUG] base_operations.py/ActiveObjectOperations.parallel_vm_config_for_ai TEST_cbuser - Parallel VM configuration for ai_1 (66E19008-7FEF-5BF4-94E8-9F366918E0A9) success.

Aug 16 14:16:24 cborch.public [2016-08-16 14:16:24,132] [DEBUG] active_operations.py/ActiveObjectOperations.objattach TEST_cbuser - AI object 66E19008-7FEF-5BF4-94E8-9F366918E0A9 (named "ai_1") sucessfully attached to this experiment. It is ssh-accessible at the IP address 10.0.0.3 (cb-cbuser-TESTPDM-vm1-iperfclient-ai-1).

15. The provisioning SLA for a given Application Instance was violated (run code = 5). How do I check which component in the creation of application instance took the most time?

The total provisioning time has 7 components, being the first an epoch timestamp and the subsequent 6 deltas (in seconds):

mgt_001_provisioning_request_originated_abs
mgt_002_provisioning_request_sent
mgt_003_provisioning_request_completed
mgt_004_network_acessible
mgt_005_file_transfer
mgt_006_instance_preparation
mgt_007_application_start

These values are all stored on the file “VM_management_*.csv”, present on the directory

~/results/<EXPID>/perf/<EXPID><AI TYPE>[BASELINE, ELASTICITY]<TIMESTAMP>. Additionally, the detailed timeline described on the previous question will also output the specific reason for a provisioning qos violation on a given Application Instance.

16. Which OpenStack releases are supported by SPEC Cloud IaaS 2016?

CBTOOL uses OpenStack APIs. If there are no changes in OpenStack APIs, new releases should continue to be supported. SPEC Cloud has been tested up to Newton release of OpenStack.

Table Of Contents

Previous topic

Submit Results

This Page