We've had several locklosses throughout the day due to the 18040 Hz ETMY PI mode.
This last lock at 25W, I finally found a combination of settings that seemed to be actively damping it according to the OMC DCPDs. (Using the QPD as the error signal, ETMY_PI_DAMP_MODE4, -60deg +30deg -3000gain). In the OMC DCPD spectrum, the peak was going down. According to the RMS monitor of this mode in the DCPDs, I was winning. However, according to the RMS monitor of this mode in the ETMY QPD, the mode was still going up. We lost lock because of this. I don't know of any DQed high freq channels for the QPDs, so we can't look back in time at the spectrum to see if that helps illuminate the situation.
Anyhow, if someone can help me understand how a mode can seem like it's going down according to the OMC DCPDs, but still actually be increasing according to the transmon QPDs and causing a lockloss, that would be great.
Title: 07/20/2016, Day Shift 23:00 – 07:00 (16:00 – 00:00) All times in UTC (PT)
State of H1: IFO locked. Wind is a Light to Gentle Breeze (4 to 12mph). Seismic activity is low, with the exception of End-Y. It is a bit elevated; but should pose no operational difficulties. Microseism is also low.
Commissioning: Commissioners are commissioning.
Outgoing Operator: Ed
Activity Log: All Times in UTC (PT)
23:00 (16:00) Take over from Ed
23:41 (16:41) Gerardo – Back from N2 overfill
Title: 07/20/2016, Day Shift 23:00 – 07:00 (16:00 – 00:00) All times in UTC (PT)
Support: Jenne, Sheila, Kiwamu
Incoming Operator: N/A
Shift Detail Summary: IFO has been up and down as the commissioner's were working through a variety issues and improvements. Did one initial alignment and several relocks. All went well. Environmental conditions remained favorable for commissioning work during the entire shift. There were 3 earthquakes during the last hour of the shift (4.8, 4.8, 5.3mag), which (per Terramon) did not represent much of a threat to the IFO lock. They did pass without any notice.
Cleared TIM errors: H1IOPSUSEY, H1SUSETMY, H1SUSETMYPI, H1SUETMXPI
Cleared IPC errors: H1IOPSEIEY, H1ISIETMY, H1PEWMEY, H1ALSEY, H1IOPSEIEX, H1PEMEX, H1ALSEX, H1SCEX
Cleared ADC errors: H1IOPSUSEX
We certainly noticed the earthquake, but we held lock at 20W throughout.
The goal of yesterdays firmware upgrade was to change the behavior of the blinking LED lights and to allow for additional diagnostics of the problem. In the past the LED typically had three states: (i) Off, (ii) Blinking and (iii) On. This upgrade interchanged the states (ii) and (iii). For a typical link light this means: Off indicates no fiber signal, Blinking indicates a fiber signal is present, and On means the timing signal is received and synchronized. Operating without error now means all LEDs are On.
Furthermore, it is possible with the Altium JTAG port to change the behavior of the LED state while running and without disrupting the timing signal. LEDs which are on can be made to blink in rates of 0.5Hz, 1Hz, 2Hz, 4Hz, etc. This should allow us to isolate the coupling site.
The upgrade was successful for all timing slaves, but failed for the timing master/fanouts. Strangely, it failed due to bogus code in svn on how to calculated the CRC-32 (ethernet) checksum. Baffling. More investigations are ongoing.
A new version 3 of the master fanout firmware is ready for test and available here. This version cleans up some timing closure issues during synthesis/translation and gerenates a correct CRC-32.
Signal for GV15 AIP went red around 18:41 utc, noticed it after lunch, went out to the mid station and it turned out to be a bad extension cord (the wire crimped to the spade slipped out of its crimped joint), replaced the bad cord with another one for now, we will install the permanent one soon.
Note: This is the 3rd cord of this type that has failed in a similar manner, we will be replacing all extension cords of this type in the near future.
Over-filled CP3 with the exhaust bypass valve fully open and the LLCV bypass valve 1/2 turn open.
Flow was noted after 45 minutes, closed LLCV valve, and 3 minutes later the exhaust bypass valve was closed.
I did a trend of the level and noticed that sometime Monday the slope of the consumption changed (it slowed down considerably). See attached.
I changed the manual setting from 17 up to 25, left such setting there for 2 hours, it quite did not look right so I dropped it to 24.
We closed some alignement loops for SRM using a dither and demodulating POP90 this morning. These loops seem to have a ugf below 100 mHz, so we may want to increase their gain, but they seem to be working well for now. They come on in the guardian durring the DRMI ASC and leave them on. They maintain POP 90 (and the ratio of AS90/POP90) at a reasonable level as we go through the CARM offset reduction, engaging the soft loops, and increasing the power.
We saw a instability that seemed to be a cross coupling between SRM and SOFT yaw loops, Jenne increased the gain in the soft yaw loops by a factor of 10 which seems to have taken care of the problem and is now in the guardian.
As long as this keeps working, operators should no longer have to adjust SRM.
Somehow, running this loop durring acquisition was causing an HLTS bounce mode (28 Hz ish, probably PR3) to ring up, which also saturated the quad L2 drives and therefore cause violins to ring up. We are now using the normal 36 WFS during DRMI, turning them off, and turning on the dither loop in the ENGAGE_SRC_ASC state once we reach full lock. This donesn't ring anything up.
Hugh, JeffK, JimW
Short version:
I noticed this afternoon that all of the BSC-ISIs were pushing large DC-ish drives on their St1 vertical actuators (generally about 2000 counts, normal drives are ~100). When I trended the BS and ITMX drives against their Sus oplevs, there seems to be a very clear correlation between large ISI drives and oplev yaw. I think this may be the cause of the drifts that Sheila complained about yesterday. The first attached plot shows the BS 3 vertical drives and BS oplev yaw, second is for the same for ITMX.
One possible cause of this is the actuators heating up the ISI. During the timing kerfluffle yesterday, all of the ISIs sat for ~6 hours with no acuator drive. The control room then recovered all of the platforms and proceeded to start aligning the interferometer, so the ISIs went from a "cold", immediately to an aligned state. Arnaud found a similar effect at LLO, where large ISI tidal drives caused angular drift during locks, when LLO tried offloading tidal to the ISI ( https://alog.ligo-la.caltech.edu/aLOG/index.php?callRep=20222 ).
In the future we should try to at least get the ISIs damped as soon as possible after similar maintenance, or go through a cycle of de-isolating/re-isolating after the ISIs have been running for a while, before trying to relock the IFO. The morning of Jul 11th, ITMX was down for almost 14hrs, but the damping loops were still on, and the actuators were still getting a drive signal. When the ISI was re-isolated, the oplev didn't see any appreciable drift over the next several hours (see third plot). Similarly, sometime before the commissioners start locking tomorrow, all the BSCs should be taken down to damped and then re-isolated.
We also considered that something in the timing system could have been responsible, but were unable to find anything to support that. None of the timing signals we looked at showed anything like the drift we saw in the ISIs. We also saw a similar alignment drift after the June power outage, that would seem to support a thermal drift over a some timing shenanigans.
I can think of two fixes for this,
1) we could have a new "off" state where the actuators have a DC value that represents the offset that they were at in the running state
2) we measure the temperature of the the stages (Quad top mass vertical position for stage two, possibly stage one vertical force for stage one) and feed forward to the yaw control
Arnaud's comment in case we come across this problem again: turn off the RZ loops. The drift is caused (we think) by thermal expansion of the stage, this changes the free hanging position of the ISI and the drift that the interferometer sees is likely dominated by RZ. If we turn the loop off, the stage will stay in a neutral position, rather than following a spurious thermally driven alignment.
As Hugh reported ealier today, the hydrolic fluid pressure as read by the HEPI pump controller units dropped yesterday when the timing FPGAs were being reprogrammed. He also reported that this was seen the last time we reprogrammed the FPGAs on Tuesday 19th May 2015.
We are seeing other unrelated EPICS channels showing data issues around the time of timing upgrades. Attached are dataviewer plots (24 hours of minute trends) which show an EY vacuum gauge and the EY-HEPI pump controller's 4th pressure sensor for 19 May 2015 and 19 July 2016. In both cases as the timing is being upgraded the fluid pressure appears to drop and the vacuum pressure appers to increase.
Investigation continues. We may ask for some EY downtime next Tuesday to see if this is reproducible and to more closely monitor what is changing. At this point we are assuming this is a sensor or read-out error and the vacuum/fluid pressures were not actually changed.
We have been studying the list of blip glitches that Miriam Cabero Müller generated for O1. We noticed that the rate of blip glitches increased dramatically during two time periods, see figure 1. We first checked if the glitch rate might be correlated with outside temperature, in case the blip glitches were produced by beam tube particulate events. The correlation with outside temperature was strong, but, for beam tube particulate, we expected it to be stronger with rate of temperature change than with temperature, and it was not. So we checked relative humidity and found that inside relative humidity was well correlated with outside temperature (and glitch rate), most likely because of the extra heating needed in cold spells. A plot of the blip glitch rate and RH inside the CS mass storage room is attached.
While the correlation with inside relative humidity is not better than with outside temperature, we plot inside relative humidity because we can better speculate on reasons for the correlation. Dry conditions may lead to the build up and discharge of static electricity on electronics cooling fans. Alternatively, there may be current leakage paths that are more likely to discharge in bursts when the pathways dry out. While this is, at best, speculation, we set up a magnetometer near the HV line for the EY ESD to monitor for possible small short fluctuations in current that are correlated with blip glitches. At the same time, we suggest that, as risk mitigation, we consider humidifying the experimental areas during cold snaps.
The low-humidity correlated blip glitches may represent a different population of glitches because they have statistically significantly smaller SNRs than the background blip glitches. We analyzed the distribution of SNR (as reported by pycbc) of the blip glitches during three time periods – segments 1, 2, and a relatively quiet period from October 5 – October 20 (segment 3). This gave approximately 600 blip glitches for each segment. Figure 2 is a histogram of these
To determine if these distributions are statistically different, we used the Mann-Whitney U test. Segments 2 and 3 matched, reporting a one-sided p-value of 0.18. The distribution in SNR for segment 3 - the low glitch rate times - did not match either segment 1 or 2, with p-values of 0.0015 and 2.0e-5, respectively. Thus we can conclude that the distributions of 1 and 2 are statistically significantly different from 3.
We are currently examining the diurnal variations in the rate of these blip glitches, and will post an alog about that soon.
Paul Schale, Robert Schofield, Jordan Palamos
This is probably something you already checked, but could it be just that there is more heating going on when the outside temperature is low? More heating would mean more energy consumption, which I guess could bother the electronics in a variety of ways that have nothing to do with humidity (magnetic fields, power glitches, vibration, acoustics in the VEAs, etc.). Which other coupling mechanisms have you investigated?
In a sense we are suggesting that the effect is due to the extra heating during the cold snaps, the humidity is just our best guess as to how the extra heating affects the electronics. We think it is unlikely to be temperature, since the temperature in the buildings changed little and did not correlate as well as humidity. Our understanding is that DetChar and others have looked carefully for, and not found, coincident events in auxiliary channels, which would argue against magnetic or power glitches from heaters. The heaters dont increase vibration or sound levels by much, the fans work continuously. The humidity, however, changed a lot.
Could the decrease in humidity affect the electronics cooling efficiency by changing the heat capacity of the cooling air? Is there any recorded direct measurement of the electronics heat-sink temperatures or exhaust temperature?
If you want to do a similar study at L1 then one of the best periods (in terms of a fairly quick change in RH) is the period from Nov. 20th to Nov. 27th 2015. Of course RH values here in the swamps are much higher. Where is the "blip glitch" list? I followed this link: https://wiki.ligo.org/viewauth/DetChar/GlitchClassificationStudy but there's nothing past Sept. 23 there.
The list of blip glitches was emailed out by Miriam Cabero Müller. They're store at https://www.atlas.aei.uni-hannover.de/~miriam.cabero/LSC/blips/full_O1/, and Omega scans for H1 are here: https://ldas-jobs.ligo-wa.caltech.edu/~miriam.cabero/blips/wscan_tables/ and for L1 here: https://ldas-jobs.ligo-la.caltech.edu/~miriam.cabero/blips/wscan_tables/.
John:
Humidity has very little effect on the thermal properties of air at the relevant temperatures and humidities (20-40 degrees C, 0-20 % RH). On pages 1104 and 1105 of this paper (http://www.ewp.rpi.edu/hartford/~roberk/IS_Climate/Impacts/Resources/calculate%20mixture%20viscosity.pdf), there are plots of specific heat capacity, viscosity, thermal conductivity, and thermal diffusivity.
TITLE: 07/20 Day Shift: 15:00-23:00 UTC (08:00-16:00 PST), all times posted in UTC
STATE of H1: Commissioning
INCOMING OPERATOR: Jeff
SHIFT SUMMARY:
LOG:
16:54 Initial Alignment
17:18 John and Bubba to My
17:30 Christina to Mid Stations for cleaning
19:29 Christina leaving MX
20:26 LockLoss due to increased PI in ETMY QPD Modes 4&6
21:02 Gerardo out to MX to investigate possible failed GV15 annulus ion pump.
21:14 Richard out to the floor
21:19 Richard out
22:16 Gerardo to MY for CP3 LN2 spillage
Ed, Jeff K, Sheila, Jenne...
at 20:26 UTC a steep rise in ETMY_PI_DAMP_MODE4_RMSMON and ETMY_PI_OMC_DAMP_MODE6_RMSMON caused the lockloss.
As was suggested a couple of days ago (28480), we recentered the POP QPDs by moving two pico motors in HAM3. The centering was done when the interferometer was fully locked with a 2W PSL and with the full ASCs engaged (except for PRC1 which uses a POP QPD).
Here are the pulse counts that we introduced to the pico-motors (pico A or controller 3) today.
After pico-motoring, we took out the offset in POP_A(B)_PIT(YAW)_OFFSET and re-closed PRC1 without an issue.
J. Kissel, E. Hall Checking in on the power delivered to the ITMY compensation plate after the strange drop in TSCY's front-end laser power drop yesterday (see LHO aLOG 28506), it looks like the laser power has recovered, mostly. The power, as measured by the pick-off beam just before the up-periscope into the vacuum system (H1:TCS-ITMY_CO2_LSRPWR_MTR_OUTPUT) is roughly stable at 0.3075 [W], where it used to deliver a really stable 0.312 [W]. I attach two zoom levels of the same 4-day trend. There's also some weird, 10-min period feature in the *minimum* of the minute-trend, where the reported power drops to 0.16 [W]. Given its periodicity, one might immediate suspect data viewer and data retrieval problems, but one can see in the un-zoomed trend that this half-power drop has been happening since the drop-out yesterday, but tracks the reported laser power even before the delivered power was increased back to nominal.
I'm wondering if this weird behaviour is due to the RF driver - we should try to swap in the spare driver soon to check this because if the power output is really glitching low like that then it's likely to cause issues for commisioning, or possibly to fail altogether.
The temperature trend for the laser doesn't give any signs that it might have overheated.
The delivered power was recovered because I added an offset to the RS calibration so that it allows the right amount of power through. The laser itself is still outputting 42W.
At 8:09 UTC (on July 20th) we had a large glitch in interferometer signals and MC3 saturated. This seems like a suspect for some kind of a suspension glitch that would be worth following up on.
The glitch is at 8:08:24.5 UTC, and it seems to have originated from the MC3 M1 LF OSEM. The glitch there is 500 counts peak-to-peak, while it's about 40 in the RT OSEM. It looks like one short glitch that caused some ringing for three seconds, which was visible in T2 and T3 as well, but not nearly as large as LF and RT.
For what it's worth - the overall trend of these signals did not change from before the little glitch event. OSEM signals look healthy.
We had a similarly huge glitch a few seconds before 23:30:30 UTC (also July 20th). It doesn't seem to be the same MC3 LF problem that Andy found. In the first attachment you can see the glitch from last night showing up clearly in MC3 M1 LF, in the second attachment you can see the very similar glitch from this afternoon without anything happening in MC3. For this afternoon's glitch I also looked at top mass osems for all the other optics and don't see much happening.
Also, all 3 MC mirrors react to the 8:08 glitch, this is because we don't have much of a cut off on our MC WFS loops. Adding more cut offs might be a good idea.
We've had several unexplained and sudden locklosses lately, and I wonder if whatever causes these huge glitches also causes some locklosses.
Replaced the analog camera looking at the BS with a GigE camera. Camera 29 on the MEDM. Ran the network cable from the camera to SUS-R2 Patch panel port 5. This is a run to the CER. In the CER plugged the unit into port 35 on the network switch. We have a view that will need to be refocused when the IFO relocks. Also have the illuminator on so we have a better view. This needs to be unplugged at the illuminator.
> Anyhow, if someone can help me understand how a mode can seem like it's going down according to the OMC DCPDs,
> but still actually be increasing according to the transmon QPDs and causing a lockloss, that would be great.
I guess there could be many explanations, but one easy one is that the dirt which produces the OMC signal might be alignment dependent. That is, for odd (n+m) optical modes the OMC transmission will depend on the OMC alignment. Since the OMC DC signal should (in a perfect world) be zero for these higher order modes, it might not be surprising that it sometimes actually approaches zero.