The turbo pump system at X-end shut down for no obvious reason around midday(?) today -> I used this "down time" do some routine maintenance on the QDP80 backing pump (grease rotor bearings, change gear oil - 400cc Krytox - and top off heat exchanger coolant but these all looked normal) -> Restarted QDP80 and MTP -> I'll leave these running but valved out from the X-end VE for overnight Operators Note: X-end under vacuum but unpumped overnight, i.e., X-end pressure will be increasing and PT510B value may change from "green" to "yellow" if left this way indefinitely - no action is required.
Got the first Corner(NE) HEPI Housing down from BSC8. The iLIGO CAS Support Stack is part way in and will continue tomorrow.
Vacuum-wipe down-vacuum complete. Inspection underway: it should be completed by the end of shift. FTIRs and installation of permanent BSC flooring tomorrow. Logistical problems trying to get a cleanroom from the cleaning area into the VEA to cover the dome. Randy is taking a look at the situation. We started hauling ICC/staging gear back to the corner.
Mark B. and Arnaud P.
Yesterday and today we made a bunch of minor model changes:
We added PCIE outputs H1:SUS-MC1_2_ISI_OFFLOAD_{X/Y/RZ/Z/RX/RY} to MC1, and similarly for MC2, MC3, PR3 (PR2 and PRM already had them).
We reconnected the watchdogs on PRM and PR3, now that they're in the chamber on the ISI.
We deleted "_2_SUS" from the names of PCIE inputs H1:ISI-ETMY_2_SUS_GS13_{X/Y/RZ/Z/RX/RY} in the TMSY model to bring them into line with what the ISI was sending. (The ETMy and ITMy models were changed to this convention earlier.) Similarly we also deleted "_ISI2SUS" from EPICS outputs (H1:SUS-)TMSY_ISI2SUS_GS13_{X/Y/RZ/Z/RX/RY} .
All of these mods were committed to the userapps SVN and were compiled and installed under rtcds 2.6.2 as part of today's upgrade.
We noted that h1susauxb123.mdl exists in the SVN but is not under version control and resolved to find out its status and get it working and/or commit it.
BS BSTS BSFM
Attached are plots of phase 2b power spectra for the beamsplitter installed in the test stand.
allbsfms_2013-02-25_H1_BS_Phase2b_ALL_Spectra_{Doff/Don}.pdf file shows comparison between LLO BS (Phase 3a Oct 2012) and LHO BSTST(Phase 2b Feb 2013).
2013-02-25_1000_H1SUSBSTST_M{1/2}_ALL_Spectra.pdf show comparison between damping on and damping off.
**Resonnances from LHO are consistent with those from LLO.
**Damping is doing his job correctly
The plots all look good, nothing to notice in particular.
Transfer functions will be running overnight to finalize Phase 2b.
PS: The script *plotallbsfm_spectra.m* has been modified, after noticing a small mistake. The pdf files comparing data from different sites were created under the name allhstss_ instead of allbsfms_.
Work Platform in place-Done Cleanroom in place-Done E-module in place (Beer garden)-Done Add hand-rail to e-module Build custom bridge between work platform and e-module Move garbing/staging cleanroom from spool area to top of e-module Move cleanroom from HAM aux optics table to BSC2-HAM3 for garbing/staging Work on cleanroom curtains-Started 26 Feb 2013 Check cleanroom function Clean chamber and cleanrooms Stage garbing/staging cleanrooms Baffle assembly (Clean test stand-solid stack to test stand ~28 Feb-Shelving units moved out-work tables moved in-clean cleanroom and contents prior to assembly start)
I turned it back on and turned off the audible alarm, as it is no longer in a clean room.
This morning I find no pressure drop from yesterday morning so I opened the second side of BSC1 and BSC2 to the pressure gauge. Maybe we'll have some info later today.
The laser tripped last night due to low flow from the diode chiller. This morning the chiller room was leaking water and I found a broken fitting inside the water tank, where water had been leaking out. I pulled the broken piece out and will install a new one hopefully tomorrow.
30mph winds all day 0800 - PSL offline; issues with diode chiller flow interlock. Michael R investigating; may need to order replacement fitting before PSL can be restarted. 0810 - Filiberto running cables in H1 electronics room and along HAM2/3. 0800-1600 Apollo crew at EY (chamber cleaning). 0930 - Corey working in LVEA eastbay. 0930 - Contractor onsite to remove old shipping pallets and crates. 1000 - Hugh in LVEA looking for HEPI issues/leaks. 1050 - Betsy working in HAM2. 1145 - Catering service onsite. 1345-1600 - Michael L and visiting scientists touring Staging, VPW and EY buildings. 1400 - Apollo installing cleanroom and work platforms over BSC2. 1547 - Mark B and Arnaud restarting multiple SUS user models after creating changes in IPC connections. No conflicts with other users expected.
Work platform assembled around BSC 2 and clean room in position over both. It is starting to look like the most likely position for the E - module is going to be in the beer garden. ICC continuing with first vacuum complete and 25% of wipe down. Generally 2-3 wipes are sufficient, which is better than I had expected.
This morning, Fil finished running the external PR3 and PRM cables from the chamber feed thru (port D3) to the R1 rack. Travis and I then connected the internal quadrapus cables from the PR3 and PRM to the D1000225 SRS cables that run down the ISI to the inside of the feedthru port. The signals all seemed to come out in the appropriate places on medm so so far, so good. We laced and stowed the cables for PR3 and then suspended the mass to verify the health of it's signals. The M2 and M3 stage OSEMs were not perfectly well aligned, so the chamberside table is not as well leveled as the ISI, causing the suspension to hang differently in the cage now. We tweeked a few top BOSEMs such that we can take TFs in prep for IAS alignment in a few weeks. (The M2 and M3 stage OSEMs don't get aligned until after that alignemt).
In order to work on PR3 I had to stand in the beam tub on the HAM3 side of the HAM2 chamber.
While there, I also:
Picked up the older witness plate that was in the beam tube near HAM2.
Moved the newly placed witness plate (laid a few days ago) into a metal holder since it was vulnerable to breakage from standing near it.
Relaced the 4 SRS cables for CB1 and CB2 (PR3 connections) up the ISI, taking care to not ground between the ISI stages. I did not have nearly enough slack to place the CB brackets in their appropriate spots on the table so I needed to pull from underneath the ISI and reclamp down the slack. The cables I moved were laced up the NE corner of the HAM2 ISI. SEI will make sure the lacing is appropriate when they test later in the month.
Added metal face shields to the PR3, MC1, MC3 - HR side of the suspensions - all are freely suspended though.
PRM has been cabled and top stage BOSEMs tuned. It has been handed over to the testers for evaluation prior to IAS alignment.
The login server for ssh access to the DTS has been changed from x1login to x1dtslogin. The x1dtslogin uses ligo.org authentication of users. To access the DTS, users still need an account to be created by the local system administrator.
The default NDS for the control room computers has been changed from h1nds0 to h1nds1. This takes effect with any new shell opened since 11:15 PST today.
Lost about a PSI over the weekend and have been unable to find any leaks since Tuesday. I have again shut all the valves I can, except those leading to BSC3, to help locate the problem area.
WP 3727, Alex, Dave, Jim
at 10:14 we shut down h1dc0 and replaced it with the new h1dc1. Jim activated a second 10GBE port on the H1 FE-DAQ switch and ran a second 10G FMM line to the DAQ rack. The new h1dc1 has three 10GBE pcie cards (compared with only two in h1dc0). The extra card allows the front ends to split their DAQ data over the two cards. Since each card has 16 MX end points, and LHO will have a total of 31 front ends, this means that we can ensure that no two front ends share an end point and interfer with each other's DAQ data on reboot.
The 10GBE PCIe card arrangement as seen from the rear of h1dc1 is
| top | IN1 | FMM-013-0001 |
| mid | OUT | FMM-007-0006 |
| lower | IN2 | FMM-013-0002 |
The pci bus mapping put the output card (connected to the Fujitsu switch) between the input cards.
The MX startup script for the front ends was modified such that the FE sent data to card IN1 or IN2 depending upon their location in the rtsystab file. We restarted all the front ends MX Streams (did not require any model restarts) and all DAQ data was back by 11:34
To test the new system we power cycled various non-dolphin and non critical front ends: h1pemmx, h1susauxb6 and h1susauxb123. During each reboot no DAQ data on any other system went bad.
Alex then upgraded the h1fw1/h1nds1 system and installed the following
Tests between the unmodified h1fw0 and the new h1fw1 systems show significant speed up in trend data retrieval. Speed up in minute trend writing quickly dissipated.
The new h1dc1 machine initially only had 12GB of RAM compared with h1dc0's 24GB. Not wanting to touch h1dc0 since that is our fall back machine if h1dc1 has issues, we removed 3*8GB DIMMs from h0broadcast0 and reduced its data cache so it could work with 12GB (6*2GB) RAM. We were then able to make h1dc1 a 24GB machine. We have asked John Z to check that the broadcaster is still functioning ok with reduced memory.
Attached are plots of dust counts > .3 microns and > .5 microns in particles per cubic foot requested from 5 PM Feb. 21 to 5 PM Feb 22. Also attached are plots of the modes to show when they were running/acquiring data. Data was taken from h1nds0. 1440.0 minutes of trend displayed
System is still sealed and I have only a few connections left to snoop. The leak is very small and may be a non-issue. I capped some drain and vent points this afternoon in case it is the valves that are passing some gas. There are also the accumulator bladders...maybe I can check for those...