Summit

This section contains site specific variations for the summit.

Resources

Bare Metal Machines

  • T&S CSCs: azar2.cp.lsst.org, azar03.cp.lsst.org

  • Auxtel illumination control machine: auxtel-ill-control.cp.lsst.org

  • LOVE: love01.cp.lsst.org

  • ATCamera (Tony Johnson): auxtel-mcm.cp.lsst.org

  • CCCamera(Tony Johnson): comcam-mcm.cp.lsst.org

  • MTCamera (Tony Johnson): lsstcam-mcm.cp.lsst.org

  • M1M3 Support cRIO (Petr Kubánek): m1m3-crio-ss.cp.lsst.org

  • M1M3 VMS cRIO (Petr Kubánek): m1m3-crio-vms.cp.lsst.org

  • M1M3 TS cRIO (Petr Kubánek):m1m3-crio-ts.cp.lsst.org

  • M2 VMS cRIO (Petr Kubánek): m2-crio-vms01.cp.lsst.org

  • VMS Data Logger: vms-data.cp.lsst.org

  • Flat FiberSpectrograph Red: flat-fiberspecred.cp.lsst.org

  • Flat FiberSpectrograph Blue: flat-fiberspecblue.cp.lsst.org

  • ESS:1 (Camera Hexapod/Rotator) Controller RPi (Wouter van Reeven): hexrot-ess01.cp.lsst.org

  • ESS:2 (M2 Hexapod) Controller RPi (Wouter van Reeven): m2hex-ess01.cp.lsst.org

  • ESS:106 (M2 Mirror): m2-ess01.cp.lsst.org

  • ESS:107 (Laser Enclosure): laser-rpi.cp.lsst.org

  • ESS:111 (Camera Inlet Humidity) Controller RPi: camera-ess01.cp.lsst.org

  • ESS:112 (M2 Humidity) Controller RPi: m2-ess02.cp.lsst.org

  • ESS:113 (M1M3 Humidity) Controller RPi: m1m3-ess01.cp.lsst.org

  • ESS:201 (Auxtel Misc) Controller RPi (Wouter van Reeven): auxtel-ess01.cp.lsst.org

  • ESS:203 (Auxtel Lightning) Controller RPi (Wouter van Reeven): auxtel-lightning01.cp.lsst.org

  • ESS:204 (Auxtel Windsonic) Controller RPi (Wouter van Reeven): auxtel-ess02.cp.lsst.org

  • ESS:307 (DIMM): dimm.cp.lsst.org

LOVE Summary View

The overall system summary state view is called ASummary State View.

Shutdown Camera Services

  • Shutdown ATCamera OCS Bridge From auxtel-mcm.cp.lsst.org run:

    sudo systemctl stop ats-ocs-bridge.service
    
  • Shutdown MTCamera OCS Bridge From lsstcam-mcm.cp.lsst.org run:

    sudo systemctl stop ocs-bridge.service
    
  • Shutdown CCCamera OCS Bridge From comcam-mcm.cp.lsst.org run:

    sudo systemctl stop comcam-ocs-bridge.service
    

Shutdown bare metal LOVE

This needs to be done from love01.lsst.org as dco user:

sudo -iu dco
./shutdown_love

Shutdown T&S Bare Metal Services

  • Handle azar03:

    sudo -iu dco
    ./shutdown_gc
    
  • Handle M1M3 cRIO:

    ssh admin@m1m3-crio-ss.cp.lsst.org
    /etc/init.d/ts-M1M3support stop
    
  • Handle M1M3 VMS cRIO:

    ssh admin@m1m3-crio-vms.cp.lsst.org
    /etc/init.d/ts-VMS stop
    
  • Handle M1M3 Thermal System cRIO:

    ssh admin@m1m3-crio-ts.cp.lsst.org
    /etc/init.d/ts-m1m3thermal stop
    
  • Handle M2 VMS cRIO:

    ssh admin@m2-crio-vms01.cp.lsst.org
    /etc/init.d/ts-VMS stop
    
  • Handle VMS Data Logger:

    ssh vms-data.cp.lsst.org
    sudo systemctl stop docker.vmslogger
    
  • Handle Auxtel illumination control:

    ssh auxtel-ill-control.cp.lsst.org
    sudo -iu dco
    ./launch_fiberspec
    
  • Handle Flat FiberSpectrograph Red:

    ssh flat-fiberspecred.cp.lsst.org
    sudo -iu dco
    ./shutdown_fiberspec
    
  • Handle Flat FiberSpectrograph Blue:

    ssh flat-fiberspecblue.cp.lsst.org
    sudo -iu dco
    ./shutdown_fiberspec
    

Interacting with Kubernetes

Commands can be executed from your own machine with kubectl and the yagan.yaml kubeconfig file. You can obtain the kubeconfig file from https://rancher.cp.lsst.org. If you don’t have access, file a Jira ticket with IT. Once you’re able to log into Rancher:

  1. Select the yagan cluster.

  2. Click the Kubeconfig File button in top-right.

  3. Near bottom of dialog, click the download link.

  4. Save the config file under your local .kube directory as yagan.yaml

  5. Point to the required cluster by doing:

    export KUBECONFIG=~/.kube/yagan.yaml
    kubectl config use-context yagan
    
  6. Ensure you are pointing to the right cluster by doing:

    kubectl config current-context
    

Update ESS Controllers

  • Updating the ESS controllers requires logging into the following machines:
    • hexrot-ess01.cp.lsst.org

    • m2hex-ess01.cp.lsst.org

    • m2-ess01.cp.lsst.org

    • m2-ess02.cp.lsst.org

    • m1m3-ess01.cp.lsst.org

    • laser-rpi.cp.lsst.org

    • camera-ess01.cp.lsst.org

    • auxtel-ess01.cp.lsst.org

    • auxtel-ess02.cp.lsst.org

    • auxtel-lightning01.cp.lsst.org

  • Most use docker-compose-ops. To stop, update and restart controllers:

    sudo -iu dco
    ./shutdown_controller
    sudo ./update_repo docker-compose-ops/ <name_of_deployment_branch>
    ./launch_controller
    
  • ESS:107 (laser-rpi.cp.lsst.org) has two containers. To stop, update and restart them:

    sudo -iu dco
    ./shutdown_ess
    ./shutdown_audiotrigger
    ./update_repo docker-compose-ops/ <name_of_deployment_branch>
    ./launch_ess
    ./launch_audiotrigger
    

Update Configuration

  • Most configurations for the different applications deployed to the Summit can be found in the Phalanx repo (lsst-sqre/phalanx). Make sure those are correct.

  • Some bare metal machine configurations also need to be updated. To do so, we use the docker-compose-admin/bin/update_repo script, which is linked into the dco user home directory. The directories to be updated are:
    • docker-compose-ops (azar2, azar03, auxtel-ill-control, flat-fiberspecred, flat-fiberspecblue)

    • LOVE-integration-tools (love01)

    • To update these machines, log into them and run:

      sudo -iu dco
      sudo ./update_repo <repo-path> <branch>
      

Startup bare metal LOVE

This needs to be done from love01. After LOVE-integration-tools has been updated:

sudo -iu dco
./launch_love

Startup Camera Services

  • Startup ATCamera OCS Bridge From auxtel-mcm.cp.lsst.org run:

    sudo systemctl start ats-ocs-bridge.service
    
  • Startup MTCamera OCS Bridge From lsstcam-mcm.cp.lsst.org run:

    sudo systemctl start ocs-bridge.service
    
  • Startup CCCamera OCS Bridge From comcam-mcm.cp.lsst.org run:

    sudo systemctl start comcam-ocs-bridge.service
    
  • Ensure bridge services are running using:

    sudo systemctl status <camera-name>-ocs-bridge.service
    
  • Transition to OFFLINE_AVAILABLE:

    ccs-shell
    ccs> set target <camera-name>-ocs-bridge
    ccs> setAvailable --withLock
    ccs> exit
    

Startup T&S Bare Metal Services

  • Handle azar03:

    sudo -iu dco
    ./launch_gc
    
  • Handle Auxtel illumination control:

    sudo -iu dco
    ./launch_gc
    
  • Handle Flat FiberSpectrograph Red:

    sudo -iu dco
    ./launch_fiberspec
    
  • Handle Flat FiberSpectrograph Blue:

    sudo -iu dco
    ./launch_fiberspec
    

Enabled CSCs

The following CSCs are configured to go into ENABLED state automatically upon launching:

  • ScriptQueue:1

  • ScriptQueue:2

  • ScriptQueue:3

  • HVAC

  • WeatherForecast

There are a few CSCs that must be put into ENABLED state before declaring an end to the deployment. These are:

  • set_summary_state.py

    data:
      - [ESS:1, ENABLED]
      - [ESS:2, ENABLED]
      - [ESS:104, ENABLED]
      - [ESS:105, ENABLED]
      - [ESS:106, ENABLED]
      - [ESS:108, ENABLED]
      - [ESS:109 ENABLED]
      - [ESS:110, ENABLED]
      - [ESS:111, ENABLED]
      - [ESS:112, ENABLED]
      - [ESS:113, ENABLED]
      - [ESS:114, ENABLED]
      - [ESS:115, ENABLED]
      - [ESS:116, ENABLED]
      - [ESS:117, ENABLED]
      - [ESS:201, ENABLED]
      - [ESS:202, ENABLED]
      - [ESS:203, ENABLED]
      - [ESS:204, ENABLED]
      - [ESS:301, ENABLED]
      - [ESS:302, ENABLED]
      - [ESS:303, ENABLED]
      - [ESS:304, ENABLED]
      - [ESS:305, ENABLED]
      - [ESS:306, ENABLED]
      - [GIS, ENABLED]
      - [Watcher, ENABLED]