Greg, TJ, Dave
In order to prevent a repeat of the error encountered last Sunday morning at 03:07 PDT where a DAQ glitch on h1susey caused an invalid flag to be latched on the latest Calibration HofT code, we have put into place a temporary system to cover us until this bug is fixed in the calibration code. Trending these CRC errors shows a very low occurence (about 8 for O2 so far), however we have put this system into place on the off chance of another event in the next week.
The system has several moving parts:
Verbal Alarms
TJ has extended the verbal alarms system to monitor the CRC error counter for h1iopsusey. If this value goes above zero, the alarm will be announced just once. After the counter goes back to zero the alarm is primed ready for the next event.
Operator Instructions
I've written full instructions in the CDS wiki OperatorInstructions
Greg is the main point of contact for the operator
Restarting the DMT calibration code
For the short term, we are asking the operator to possibly restart the calibration hoft code by having open login sessions on the machines' consoles in the MSR. The instructions give details on how these are to be used.
Once the calibration code is fixed (hopefully next Tuesday 4/25) this system will be dismantled.
TITLE: 04/19 Eve Shift: 23:00-07:00 UTC (16:00-00:00 PST), all times posted in UTC STATE of H1: Observing at 62Mpc OUTGOING OPERATOR: Nutsinee CURRENT ENVIRONMENT: Wind: 6mph Gusts, 4mph 5min avg Primary useism: 0.02 μm/s Secondary useism: 0.21 μm/s QUICK SUMMARY: Recently back to observing after scheduled commissioning. No known issues.
TITLE: 04/19 Day Shift: 15:00-23:00 UTC (08:00-16:00 PST), all times posted in UTC
STATE of H1: Observing at 56Mpc
INCOMING OPERATOR: Patrick
SHIFT SUMMARY: Commissioning since morning. Recently back to Observe. No issue to report.
LOG:
15:55 Christina opening receiving roll up door
16:00 Commissioning started. Sheila started OMC dither. Chandra to LVEA
16:08 Chandra out
16:24 Robert to LVEA
18:14 Robert done/Jenne started
18:30 Hand off to Jeff/Lockloss
19:00 Back to NLN/ran a2l/ back to commissioning/ Robert to LVEA
19:13 Karen driving to VPW
19:30 Robert done. Sheila took over
19:40 Karen back
20:06 Jeff took over
20:14 Lockloss/ Jeff done
20:53 Back to NLN/ Robert to LVEA
21:33 Robert to LVEA again (when did you come back??)
22:03 Lockloss (Robert claimed it wasn't him)/ LLO called -- they're ready to come back up.
22:16 Robert out
22:48 Back to Observe
23:59 Out of Observing to let Robert go turn off a computer in the LVEA
23:03 Back to Observe
We will have another commissioning window on Thursday coincident with LLO.
The main task will be either laser vibrometer measurement of the baffles (again) or air flow test in the PSL room.
There are lower prioritiy tasks that I don't list here.
Betsy, Gerardo
Today, I assisted Gerardo in his supurb silicate bonding of ear #182 to the S4 flat of spare ITM test mass ITM06. No bubbles and placement on flat within spec with alot of headroom. The second ear is scheduled for bonding next week.
J. Kissel I attempted to repeat Evan Hall's study (LHO aLOG 27675) in which he was able to manipulate the detuning of the DARM response with a SRCL offset, but was unsuccessful. (1) I was able only to inject an offset of +/-100 [ct] in H1:LSC-SRCL1_OFFSET and have the IFO stable (I tried offsets of +/-200 like Evan had done, and then +/-150, but both cause a lock-loss) (2) With a +/-100 [ct] SRCL offset, the new SRC detuning parameter tracking did not show any signs of spring frequency change. Offset ramping times were 120 [sec]. When at +/-150 [ct], there was visible excess noise in SRCL, which then propagated through standard coupling to DARM (see attachement for comparison between no offset and -150 ct gathered just before lock loss). I did not have time to take sweeps as Evan did. Recall that at Evan's study was done at with an O1, 10 W IFO. Perhaps the calibration of the offset (he suggests 0.1 [ct/pm]) changes with input power; he doesn't give enough detail about his SRCL budget to re-derive the calibration. Will look into the theory more before attempting this again. It's discouraging that, although the SRC detuning frequency calculation from 7.93 Hz PCALY line appears roughly correct compared to recent sensing function sweeps, the calculation appears to be insensitive to this known thing that manipulates it. #backtothedrawingboard
J. Kissel, K. Izumi Izumi-sensei has come to my rescue. He imparts the following knowledge: (1) The SRCL calibration is independent of input power, and the same at DC as it is at AC because the optical plant is flat for SRCL. (2) That calibration is stored in the CAL-CS model (specifically the H1:CAL-CS_SUM_SRCL_ERR filter bank, FM1), calibrating the very channels I'd posted in the spectra above: -9.55e-6 [um/ct] = 9.55e-12 [m/ct] ~= 10 [pm/ct] == >> 0.1 [ct/pm] i.e. the number that Evan quotes in LHO aLOG 27675 which comes from Kiwamu's original model of the SRCL open loop gain (LHO aLOG 18742) (3) The proof is in the pudding: we don't scale the SRCL gain at all during a normal lock acquisition sequence when we increase from 10 to 30 W. In summary, there's no obvious reason why the OFFSET calibration should have changed between Evan's measurement and my own today. And to repeat what was said above -- there was no evidence of change in the live calculation of the SRC detuning spring frequency. Something is rotten in the state of Denmark...
R. Savage, S. Karki, T. Shaffer, J. Kissel At the request of Rick and Sudarshan, I modified the HIGH_FREQ_LINES guardian to start a new long-duration PCAL X excitation sweep to investigate an apparent systematic error they've found at the 5 kHz data point of the previous sweep (see attachment for current status of model vs. measurement, originally from here; the data from previous iterations of the guardian node code in LHO aLOGs 35532 and 35153). The new bounds of the sweep are from 4825 Hz, in 75 Hz increments up to 6 kHz, and the sweep direction is *up* instead of down. I've committed the changes to the code to userapps repo here, /opt/rtcds/userapps/release/cal/h1/guardian/HIGH_FREQ_LINES.py reloaded the guardian code, changed to the start frequency by hand, and ran the INIT to reset the 24 hr timer. The next observation stretch (after Apr 19 2017 20:25 UTC) should start the sweep and have the frequency set to 4825 Hz.
Starting CP3 fill. LLCV enabled. LLCV set to manual control. LLCV set to 50% open. Fill completed in 1230 seconds. TC B did not register fill. LLCV set back to 21.0% open. Starting CP4 fill. LLCV enabled. LLCV set to manual control. LLCV set to 70% open. Fill completed in 2061 seconds. LLCV set back to 39.0% open.
Raised each by1%
CP3 now 22% open
CP4 now 40% open
model restarts logged for Tue 18/Apr/2017
2017_04_18 10:17 h1calcs
2017_04_18 10:23 h1broadcast0
2017_04_18 10:23 h1dc0
2017_04_18 10:23 h1fw0
2017_04_18 10:23 h1fw1
2017_04_18 10:23 h1fw2
2017_04_18 10:23 h1nds0
2017_04_18 10:23 h1nds1
2017_04_18 10:23 h1tw1
Maintenance day. h1calcs model change, new H1EDCU_RACCESS.ini, DAQ restart to support these.
model restarts logged for Mon 17/Apr/2017 - Wed 12/Apr/2017 No restarts reported
WP6572 h1calcs model change
Jeff K, Dave:
new code was installed for h1calcs.
h1hwsex Recovery
Nutsinee, Dave:
The HWS code was recovered on h1hwsex after the frame grabber card was returned last week.
DMT Calibration code reversion
Greg:
The new calibration code (which was installed on 4/11) was reverted back to the old version, but this failed due to channel list changes. The new code was left in place.
Remote access channel lists updated
Jonathan, Dave:
The RACCESS channels lists in both the autoBurt.req and H1EDCU_RACCESS.ini files were updated.
DAQ restart
Dave:
DAQ was restarted to cover h1calcs and H1EDCU_RACCESS config changes
Work Permit | Date | Description | alog/status | |
|
4/18/2017 9:53 | Remove the HAM4 South door illuminator from port and look into chamber to view HWS optics. There is a clear guillotine cover in place which will not be removed. If it needs to be removed for any reason (to aid viewing) I will ask Gerardo for assistance. No laser beams are traced to be exiting this port. | 35655 | |
|
4/17/2017 15:52 | Duplicate of WP 6586 | 35624 | |
6586.html | 4/17/2017 15:52 | Remove and replace LN2 level gauge LI105A at storage tank 8514380 at the X-end*Will result in CDS level indication to read near zero (RED) for this tank while replacement is in progress | 35624 | |
|
4/17/2017 15:46 | Run cables from BRS (Beam Rotation Sensor) ion pump controller to vacuum racks for both end stations. Ion pump controllers will be elevated off the ground. | ||
|
4/17/2017 14:37 | Run new cabling for temperature sensors from the mechanical room into the LVEA. Runs in the LVEA are to the following chambers: HAM2, HAM3, HAM4, HAM5, BSC1, and BSC3. Work will require climbing on some of the chambers. This work is part of the HVAC upgrade being done by Apollo. | 35643 | |
|
4/17/2017 13:04 | Revert calibration code running on the DMT from gstlal-calibration-1.1.5 to gstlal-calibration-1.1.4. This will change the DMT calibration code to the older code, while gstlal-calibration-1.1.6 undergoes further testing. 1. This will removed the bug in gstlal-calibration-1.1.5 which causes the calibration to be marked as NOT OK persistently after data zeros or data dropout occur, e.g., as seen here: https://ldas-jobs.ligo-wa.caltech.edu/~detchar/summary/day/20170416/plots/H1-MULTI_D80927_SEGMENTS-1176336018-86400.png 2. This will re-introduce the bug in gstlal-calibration-1.1.4 which can cause the calibration to be incorrectly marked OK for -/+ 3 s after a data dropout. This problem is rare and can be vetoed later. The plan is to continue to test gstlal-calibration-1.1.6 (or later as needed) to verify it fixes both bugs listed above (and doesn't introduce new bugs) A new WP will be issued then. And C01 will be run with the new code over times since March 10, to fix the times affected by bugs in online running since then. | 35647, 35647, 35612, 35619 | |
|
4/17/2017 12:25 | Repeat tests described in LLO alog 28979 Should take less than 1 hour | ||
6581.html | 4/17/2017 12:13 | Replace razor dump on ISCT1 with black glass diamond, to reduce upconversion and DARM noise during high ground motion times. | 35636 | |
|
4/17/2017 11:37 | Remove view port cover and setup/use laser vibrometer to investigate baffles in input arm. | ||
|
4/17/2017 10:50 | Run QDP80 roughing pump in corner station mech. room, along with Kobelco (needed for purge air in roughing line during shut down). Also spin up y-beam manifold turbo pump. Swap out battery packs on X&Y beam manifold TPs. | 35625 | |
|
4/17/2017 10:00 | Regular PSL maintenance: Increase the HPO pump diode currents to return HPO output power to its start-of-O2 level. HPO power is down due to natural decay of the HPO pump diodes. Tweak beam alignment into PSL PMC to return to start-of-O2 PMC transmitted power. PMC transmitted power is down due to HPO pump diode decay and alignment shift into the PMC. This work requires the PSL ISS to be OFF, so the IFO will need to be unlocked. All work will be preformed from the control room, no PSL incursion is necessary. | 35611 | |
|
4/14/2017 11:39 | Remove the channels that have already been migrated to BACNet during the FMCS upgrade from the existing FMCS IOC where they are now invalid. Start and run a separate BACNet IOC for these channels and additional ones created during the upgrade. Continue this process as the upgrade continues. | ||
|
4/14/2017 8:53 | Change the power settings so the screen goes to sleep after 0.5 hour of inactivity. | ||
|
4/13/2017 23:26 | Perform bi-monthly PCal calibration at End X. Will require aligned ETMx and End X VEA to be laser hazard. | 35639 | |
|
4/13/2017 10:10 | Test EPICS IOC for FMCS BacNet channels. Will create a number of EPICS channels on the CDS network. | ||
Previous W.P. | ||||
6572.html | 04/11/17 12:41 PM | Install SRC detuning time dependent tracking. Includes update to CAL-CS front-end model, GDS gst-lal calibration pipeline code, updates to the science frame channel list which need to be included in all down stream channel lists so that the GDS pipeline can use those channels. Requires DAQ restart, and modification to separate DMT Frame Broadcaster channel list. Corresponds to gstlal release 1.1.5. | 35462, 35474, 35478, 35637 | |
6541.html | 03/27/17 08:40 AM | Perform scheduled maintenance to scroll compressors at Y-MID vent/purge-air supply skid. Maintenance activity will require for the compressors to run for brief periods of time to check compression. Lock-out/tag-out power to skid as required. | 35149, 35615 |
HWS code was writing HWSY data overnight. Stopped at 16:18:15 UTC
Swapped camera fiber cable 16:19 UTC
Code restarted 16:20 UTC, now writing HWSX data.
See alog35628 for details
16:00-16:07 UTC
entered LVEA to de-energize three turbo stations (levitated bearing in turbo dropped)
TITLE: 04/19 Owl Shift: 07:00-15:00 UTC (00:00-08:00 PST), all times posted in UTC
STATE of H1: Observing at 67Mpc
INCOMING OPERATOR: Nutsinee
SHIFT SUMMARY:
Other than a lockloss, mostly uneventful night mostly in OBSERVING.
LOG:
In order to see if a couple of optics for HWSY could be identified, the illuminator cover on the A1F4 viewport of HAM4 was removed. Conclusion The optic in the HWSY DCBS1 lens mount is most likely to be a dichroic beamsplitter rather than a lens. The orientation of the HWSY VacLens mount is such that the retaining ring is facing the HWS table. Clearly seen are the two holes of the retaining ring as Betsy illuminated the chamber from A1F4. The holes are not clearly visible when one looks from the A1F4 flange side, which is consistent with them being partly blocked by the lens mount. The 2" Siskiyou lens mount is 0.50" thick but has a 0.41" deep recess to accommodate an optic. The retaining ring for the mount is 0.13" thick. If the optic in the mount is the dichroic beamsplitter (D1200214) which is 0.375" thick. Since the stack up of the retaining ring and a 3/8" thick optic is thicker than the mount, the mount is a custom modification to the standard part. More importantly since the retaining ring appears to be almost flush with the surface of the lens mount, the optic held is more likely to be the dichroic beamsplitter rather than a lens. The data sheet for a PLCX-50.8-360.6 lens says it is 6.4 mm thick - a concave lens would be similar. If the optic in the mount were a lens, the retaining ring would sit deeper in the mount. Betsy / Peter
The modified Siskiyou lens mount dimensions state that it is made from a standard part. Something doesn't quite hang together though because it's not clear to me how one makes a 0.56" deep recess with a 0.50" thick part. Unless either 0.56" is really 0.46", or the part is not 0.50" thick as per the vendor drawing. /* update */ Eddie Sanchez provided me with this information. D1102166 is 0.65" thick, not 0.50". The up shot is that the optic concerned is still more likely to be the dichroic beamsplitter and not a lens.
If I recall correctly this image (attached) is normally dark in NLN?
Looking at the auto-captured images, indeed you are correct that there is some light where there didn't used to be any.
The work that could have affected this is the placement of the new black glass beam dump (alog 35636). Not that there should be any significant reflection from this dump (it has 5 bounces from black glass), any reflection would go in the opposite direction from the ALSX camera. So, I don't think it could be reflection from this beam dump. What is possible is that the old razor dump was catching stray beams in addition to the main one, and the black glass dump with a slightly smaller profile isn't catching that stray beam.
I think dumping the main POP beam with the new dump is still overall helpful, and we'll see if this extra stray beam actually has any affect on the interferometer. It's small enough in power that it wasn't visible on the IR card we were using, although we didn't utilize an IR viewer to see much dimmer beams.
Right now, I think we should monitor things to see if this beam has any effect, and we can look at ISCT1 again next Tuesday.
Evan G., Robert S. Looking back at Keith R.'s aLOGs documenting a changes happening on March 14 (see 35146, 35274, and 35328), we found that one cause seems to be the shuttering of the OpLev lasers on March 14. Right around this time, 17:00 UTC on March 14 at EY and 16:07 UTC at EX, there is an increase in line activity. The correlated cause is Travis' visit to the end station to take images of the Pcal spot positions. The images are taken using the Pcal camera system and needs the OpLevs to be shuttered so that a clean image can be taken without the light contamination. We spoke with Travis and he explained that he disconnected the USB interface between the DSLR and the ethernet adapter, and used a laptop to directly take images. Around this time, the lines seem to get worse in the magnetometer channels (see, for example, the plots attached to Keith's aLOG 35328). After establishing this connection, we went to the end stations to turn off the ethernet adapters for the Pcal cameras (the cameras are blocked anyway, so this active connection is not needed). I made some magnetometer spectra before and after this change (see attached). This shows that a number of lines in the magnetometers are reduced or are now down in the noise. Hopefully this will mitigate some of the recent reports of combs in h(t). We also performed a short test turning off another ethernet adapter for the H1 illuminator and PD. This was turned off at 20:05:16 18/04/2014 UTC and turned back on at 20:09:56 UTC. I'll post another aLOG with this investigation as well.
Good work! That did a lot of good in DARM. Attached are spectra in which many narrow lines went away or were reduced (comparing 22 hours of FScan SFTs before the change (Apr 18) with 10 hours of SFTs after the change (Apr 19). We will need to collect much more data to verify that all of the degradation that began March 14 has been mitigated, but this first look is very promising - many thanks! Fig 1: 20-50 Hz Fig 2: 50-100 Hz Fig 3: 100-200 Hz
Attached are post-change spectra using another 15 hours of FScan SFTs since yesterday. Things continue to look good. Fig 1: 20-50 Hz Fig 2: 50-100 Hz Fig 3: 100-200 Hz
Correction: the date is 18/04/2017 UTC.
Another follow-up with more statistics. The mitigation from turning off the ethernet adapter continues to be confirmed with greater certainty. Figures 1-3 show spectra from pre-March 14 (1210 hours), a sample of post-March 14 data (242 hours) and post-April 18 (157 hours) for 20-50 Hz, 50-100 Hz and 100-200 Hz. With enough post-April 18 statistics, one can also look more closely at the difference between pre-March 14 and and post-April 18. Figures 4-6 and 7-9 show such comparisons with different orderings and threrefore different overlays of the curves. It appears there are lines in the post-April 18 data that are stronger than in the pre-March 14 data and lines in the earlier data that are not present in the recent data. Most notably, 1-Hz combs with +0.25-Hz and 0.50-Hz offsets from integers have disappeared. Narrow low-frequency lines that are distinctly stronger in recent data include these frequencies: 21.4286 Hz 22.7882 Hz - splitting of 0.0468 Hz 27.4170 Hz 28.214 Hz 28.6100 Hz - PEM in O1 31.4127 Hz and 2nd harmonic at 62.8254 Hz 34.1840 Hz 34.909 Hz (absent in earlier data) 41.8833 Hz 43.409 Hz (absent in earlier data) 43.919 Hz 45.579 Hz 46.9496 Hz 47.6833 Hz 56.9730 Hz 57.5889 Hz 66.7502 Hz (part of 1 Hz comb in O1) 68.3677 Hz 79.763 Hz 83.315 Hz 83.335 Hz 85.7139 Hz 85.8298 Hz 88.8895 Hz 91.158 Hz 93.8995 Hz 95.995 Hz (absent in earlier data) 107.1182 Hz 114.000 Hz (absent in earlier data) Narrow low-frequency lines in the earlier data that no longer appear include these frequencies: 20.25 Hz - 50.25 Hz (1-Hz comb wiped out!) 24.50 Hz - 62.50 Hz (1-Hz comb wiped out!) 29.1957 Hz 29.969 Hz Note that I'm not claiming change points occurred for the above lines on March 14 (as I did for the original set of lines flagged) or on April 18. I'm merely noting a difference in average line strengths before March 14 vs after April 18. Change points could have occurred between March 14 and April 18, shortly before March 14 or shortly after April 18.
To pin down better when the two 1-Hz combs disappeared from DARM, I checked Ansel's handy-dandy comb tracker and found the answer immediately. The two attached figures (screen grabs) show the summed power in the teeth of those combs. The 0.5-Hz offset comb is elevated before March 14, jumps up after March 14 and drops down to normal after April 18. The 0.25-Hz offset comb is highly elevated before March 14, jumps way up after March 14 and drops down to normal after April 18. These plots raise the interesting question of what was done on April 18 that went beyond the mitigation of the problems triggered on March 14. Figure 1 - Strength of 1-Hz comb (0.5-Hz offset) vs time (March 14 is day 547 after 9/15/2014, April 18 is day 582) Figure 2 - Strength of 1-Hz comb (0.25-Hz offset) vs time