|
|
(25 intermediate revisions by 2 users not shown) |
Line 1: |
Line 1: |
− | ==PRELIMINARY / UNDER CONSTRUCTION==
| + | '''<span style="color: red">This tutorial page needs updating.</span>''' |
− | This page describes how to run the CSIT test suites on a Mac OS X (El Capitan / 10.11.x).
| + | |
− | | + | |
− | ===Prerequisites===
| + | |
− | This procedure requires the following software be installed on the host PRIOR to following the directions:
| + | |
− | * pip (sudo easy_install pip)
| + | |
− | * virtualenv (pip install virtualenv)
| + | |
− | * [http://brew.sh Homebrew]
| + | |
− | * [https://git-scm.com/download/mac git]
| + | |
− | * ssh-copy-id (brew install ssh-copy-id)
| + | |
− | * [https://www.virtualbox.org/wiki/Downloads Virtualbox 5.x or greater]
| + | |
− | * [https://www.vagrantup.com/downloads.html Vagrant]
| + | |
− | * vagrant-cachier (vagrant plugin install vagrant-cachier)
| + | |
− | * [https://gerrit.fd.io/r/#/admin/projects/csit CSIT project source code]
| + | |
− | Optional
| + | |
− | * [https://gerrit.fd.io/r/#/admin/projects/vpp VPP project source code]
| + | |
− | The following wiki page describes how to build VPP in a VM locally:
| + | |
− | [[VPP/Pulling,_Building,_Running,_Hacking_and_Pushing_VPP_Code| Pulling, Building, Hacking, and Pushing VPP Code]]
| + | |
− | | + | |
− | | + | |
− | This guide was verified on a Mac OS X MacBook Pro /w 16GB ram running Mac OS X 10.11.5 (El Capitan). In addition to the host OS, there are 4 Virtualbox VM's required for test development: 1 VM for building VPP (vpp) and 3 VM's used by CSIT to run the test suites. The standard 3 node CSIT topology consists of a traffic generator (tg) and two Device-Under-Test machines (dut1 and dut2).
| + | |
− | | + | |
− | This guide goes through setup of the management network that is interconnecting all VMs. This network is used by the test framework to connect to topology nodes and execute test code on them. Furthermore it explains how to start the Vagrant VMs, how to install prepared deb packages on DUTs, and how to start tests.
| + | |
− | | + | |
− | ===Start Vagrant VM environment===
| + | |
− | Run the "Terminal" application to get access to the bash command line prompt. Create a folder where your Vagrant environment is going to exist, and copy the Vagrantfile and install_debs.sh from ${CSIT_DIR}/resources/tools/vagrant into that folder. Also copy all of the vpp packages to be tested into the same folder.
| + | |
− | | + | |
− | <pre>mkdir csit-vagrant
| + | |
− | cd csit-vagrant
| + | |
− | cp ${CSIT-DIR}/resources/tools/vagrant/* .
| + | |
− | cp ${VPP-DIR}/build-root/vpp*.deb .
| + | |
− | vagrant up --parallel --provision
| + | |
− | </pre>
| + | |
− | <pre>Bringing machine 'tg' up with 'virtualbox' provider...
| + | |
− | Bringing machine 'dut1' up with 'virtualbox' provider...
| + | |
− | Bringing machine 'dut2' up with 'virtualbox' provider...
| + | |
− | ...
| + | |
− | </pre> | + | |
− | | + | |
− | Vagrant will download base disk images and apply scripts to bring those machines to required state.
| + | |
− | | + | |
− | ===Copy your ssh-key to Vagrant VMs=== | + | |
− | This steps has to be repeated every time your Vagrant VMs are re-created (i.e. vagrant destroy command was issued)
| + | |
− | <pre>$ echo csit@192.168.255.10{0,1,2} | xargs -n 1 ssh-copy-id </pre>
| + | |
− | | + | |
− | Respond with "csit" as password (without quotes). From now on you have password-less access from this account to csit@vagrant-vms via SSH.
| + | |
− | | + | |
− | ===Install vpp installation packages on DUTs===
| + | |
− | To test anything you have to install the debian packages VPP build produces. If you don't have any handy, go and download latest ones from [http://nexus.fd.io].
| + | |
− | | + | |
− | Copy your packages to some location on your dev machine, and issue this command:
| + | |
− | <pre>$ resources/tools/vagrant/install_debs.sh ${DIRECTORY_WITH_YOUR_VPP_PACKAGES}/vpp*.deb</pre>
| + | |
− | | + | |
− | Pay attention to the last line of the previous command, if everything went as it was supposed to, you'll see "Success!", and exit status will be 0.
| + | |
− | | + | |
− | ===Set up your virtualenv===
| + | |
− | <pre>cd ${CSIT_DIR}
| + | |
− | rm -rf env
| + | |
− | virtualenv env
| + | |
− | source env/bin/activate
| + | |
− | pip install -r requirements.txt
| + | |
− | </pre>
| + | |
− | | + | |
− | You should now see (env) in front of your bash prompt.
| + | |
− | | + | |
− | ===Create topology file===
| + | |
− | CSIT framework uses YAML format to describe the nodes of the topology the testcases are going to run on. There are such data as IP addresses, login information, type of the node in the topology file. CSIT framework uses PCI addresses of the NICs on the topology nodes to map them to the node topology information. Luckily, PCI addresses stay at constant values in between "Vagrant up" cycles, therefore the PCI addresses are pre-stored for you in topologies/available/vagrant.yaml. BUT, within the test cases, and concretely in code that matches topology interfaces against VPP reported interfaces, MAC addreses are used. These are different every time you create new Vagrant instances of those VMs, therefore you have to scrape the PCI_ADDRESS to MAC_ADDRESS map from current topology instance. TL;DR: you have to update your topology file with MAC addresses from current running VMs.
| + | |
− | | + | |
− | This is currently automatized in some essence by running this command line: | + | |
− | <pre>(env)username@hostname:${CSIT_DIR}$ cd ${CSIT_DIR}
| + | |
− | (env)username@hostname:${CSIT_DIR}$ export PYTHONPATH=`pwd`
| + | |
− | (env)username@hostname:${CSIT_DIR}$ ./resources/tools/topology/update_topology.py -f -v -o topologies/available/vagrant_pci.yaml topologies/available/vagrant.yaml
| + | |
− | 192.168.255.101: Found MAC address of PCI device 0000:00:0a.0: 08:00:27:66:a5:75
| + | |
− | 192.168.255.101: Found MAC address of PCI device 0000:00:09.0: 08:00:27:bf:ed:90
| + | |
− | 192.168.255.100: Found MAC address of PCI device 0000:00:0a.0: 08:00:27:ae:26:e9
| + | |
− | 192.168.255.100: Found MAC address of PCI device 0000:00:09.0: 08:00:27:50:cf:7e
| + | |
− | 192.168.255.102: Found MAC address of PCI device 0000:00:0a.0: 08:00:27:2c:25:6e
| + | |
− | 192.168.255.102: Found MAC address of PCI device 0000:00:09.0: 08:00:27:41:45:7d
| + | |
− | (env)username@hostname:${CSIT_DIR}$ echo $? # this should print 0
| + | |
− | </pre> | + | |
− | | + | |
− | ===Executing test cases for Vagrant setup===
| + | |
− | <pre>
| + | |
− | (env)username@hostname:${CSIT_DIR}$ cd ${CSIT_DIR}
| + | |
− | (env)username@hostname:${CSIT_DIR}$ export PYTHONPATH=`pwd`
| + | |
− | (env)username@hostname:${CSIT_DIR}$ pybot -L TRACE -v TOPOLOGY_PATH:topologies/available/vagrant_pci.yaml --exclude 3_node_double_link_topoNOT3_node_single_link_topo --include VM_ENV --exclude PERFTEST tests/</pre>
| + | |
− | This command executes tests with TRACE logging enabled, on the topology you just updated with proper MAC addresses, and starts only testcases that are made for single-link-topology (what we have in Vagrant currently), that are made for VM environment, and are not performance tests.
| + | |
− | | + | |
− | One can modify the above command to start for example only ipv4 tests:
| + | |
− | <pre>(env)username@hostname:${CSIT_DIR}$ pybot -L TRACE -v TOPOLOGY_PATH:topologies/available/vagrant_pci.yaml -s ipv4 tests/</pre>
| + | |