I'm starting to look at the nice data set Camilla took with different non linear gains in the interferometer, 83370. A few weeks ago we took a similar dataset on the homodyne, see 83040.
OPO threshold and NLG measurements by both methods now make sense:
Based on the important realization in 83032 that we previously had pump depletion while we were measuring NLG by injecting a seed beam, we reduced the seed power for these two more recent datasets. This means that we can rely on the NLG measurements to fit the OPO threshold, and not have the OPO threshold as a free parameter in this dataset. (In the homodyne dataset, I had the OPO threshold as a free parameter, and it fit the NLG data well. With the IFO data, we want to fit more parameters so it's nice to be able to fit the threshold independently.). The first attachment shows a plot of nonlinear gain measured two ways, from Camilla's table in 83370. The first method of NLG calculation is the one we normally use at LHO, where we measure the amplified seed while scanning the seed PZT, and then block the green and scan the OPO to measure the unamplified seed level (blue dots in 1st attachment). The second method is to measure the amplified and deamplified seed while the seed PZT is scanning, (max and min), and nlg = {[1+sqrt(amplified./deamplified)]/2}^2 (orange pluses in attached plot). The fit amplified/ unamplified method gives a threshold of 158.1uW OPO transmitted power in this case, while the amplified/ deamplified method gives 157/5uW.
Mean squeezing lump and estimate of eta:
The second attachment here shows all the spectra that Camilla saved in 83370. As she mentioned there, there is something strange happening in the mean squeezing spectra from 300-400 Hz, which is probably due to the ADF at 322Hz was on while the LO loop was unlocked. In the future it would be nice to turn off the ADF when we do mean squeezing so that we don't see this. This could also be adding noise at low frequencies, making the mean squeezing measurement confusing.
Mean squeezing is injecting squeezing with the LO loop unlocked, which means that it averages over the squeezing angles, and if we know the nonlinear gain (the generated squeezing) the mean squeezing level is determined only by the total efficiency. With x = sqrt(P/P_thresh)
Lockloss @ 19:54 UTC - link to lockloss tool
Possibly due to wind; lockloss tool has 'WINDY' tag and gusts have reached almost up to 30mph in the past hour.
I've added some code to the OPLEV charge measurement scripts to export the plotted data (weighted and unweighted mean charge with their associated STD and VAR), it ended up being ~10 years of measurements for each ETM. The analysis scripts both live at /ligo/svncommon/SusSVN/sus/trunk/QUAD/Common/Scripts. The saved mat files will be in the same directory as the scripts, named 'ETMX_OPLEV_CHRG.mat' & 'ETMY_OPLEV_CHRG.mat'. The data is also attached to this alog.
The data is formatted as a matlab table with each variable labeled, **it's important to note that each column (4 per variable for ETMX, 3 for ETMY) of each variable coresponds to a quadrant in the following order UL LL UR LR. Also ETMY does not have LR**
For ETMX the data range is 11/04/14 to present. The analysis script for this quad is "Long_Trend.mat"
For ETMY the data range is 01/06/15 to present. The analysis script for this quad is "Long_Trend_ETMY_LHO.mat"
If someone wants to rerun this at a later time with new measurements, you'd just have to adjust the time range at the top of the script to what you want, then go to the end of the script and change the "export_data" variable from 0 to 1.
We started to see noise peaking up again at EX so we swapped Fans to see if this is the coupling mechanism.
H1 back to observing at 18:44 UTC after a commissioning-induced lockloss and a straightforward relock.
Before observing, I accepted a few SQZ ADF servo-related SDF diffs (the servo remains off); see attached.
Noted that there was a out of nominal rise on the Y-Arm pressure plot (see first attachment), handy tool, so I started looking into it. Noted that the temperature in the VEA for the Y mid station has gone up from an nice oscillation of 63~65 °F up to a rising temperature, currently at 68 °F, maybe we are adding heat to the Y-Mid??, all pressure gauges, including the annuli ones noted the pressure rise. No attention required, but it would be nice if the temperature stopped at a nominal value.
RyanS reported getting HAM3 CPS noise notifications on DIAG_MAIN. Looking at the log, there have been several periods where the blrms noise monitors have been going over threshold. Digging in to HAM3's performance this morning, CPS noise seems to be affection mostly RZ, but the other horizontal dofs are affected. as well. First attached image, is comparision with HAM2. Top asds are the rotational dofs, bottom are the X/Y/Z dofs. HAM3 RX/RY/Z vertical dofs are all mostly similar to HAM2, but the cyan trace vs the pink trace on the top plot shows HAM3 RZ is moving much more that HAM2 RZ.
Second image shows asds of the individual sensors, the blue trace on the top plot is H2 CPS, which has a higher noise floor than the other sensors, so I suspect this sensor is the root cause of the HAM3s problems. All of the other senors seem more or less healthy. If I get a chance or the ISI becomes unstable, I would like to go to the floor or CER to power cycle some stuff. Otherwise, I can wait till tomorrow.
This trend shows when the H2 CPS started misbehaving. Top two plots are the 1-3hz log blrms for the RZ GS13s HAM2 is pretty steady, but something happens at ~10:30am pst on the 18th. Bottom plot shows the horizontal CPS during this time, H1 and H3 stay more or less the same, but H2 shows an increase in noise that has persisted until now.
Got permission from Sheila to try addressing this because locking is not going well. Powered off the CPS racks for ~10 secs and noise was gone when the signals came back up. Will keep monitoring.
Jennie W, Camilla
Today during commissioning at 16:29:41 UTC I ran the SCAN_SQZANG_FDS guardian state and it changed the squeeze angle from 170.94 degrees to 163.08 degrees.
Then Camilla and I changed the filter cavity detuning from the nominal value of -32 Hz, measured DARM between 20 to 30 Hz atr each step. This was following this entry by Vicky and Camilla. The lowest value we tried was -44 Hz and the highest was -26 Hz. The optimal seemed to be between -38 Hz and -32 Hz, eventually we chose -36 Hz as it was better 30 and 46 Hz, and among the best between 60 and 70 Hz. Above 70 Hz it is hard to see the difference between traces.
We accepted the new value for H1:IOP-LSC0_RLF_FREQ_OFS in the OBSERVE.snap file in sdf.
Mon Mar 24 10:10:42 2025 INFO: Fill completed in 10min 38secs
Gerardo confirmed a good fill curbside, audio only, no line to sight due to tumbleweed.
J. Kissel One step closer to driving an analog signal into the OMC DCPD test chain at ~10 kHz (see other prep aLOG discussion in LHO:83466 and LHO:83468), I set up the SR785, SR785 accessory box, and at the necessary cabling by ISC-R5 rack on the -Y side of the -Y side of HAM6, tucked up against the mini clean room that's there. Other commissioning activities broke the lock just as I was about to plug the setup into the wall and power on. So -- we'll try again Thursday 03/27. All in-rack cabling is the same as is typically is for all of O4, I hadn't gotten to disconnecting the ISC_444 injection port cable yet. Was in the LVEA for a mere 6 minutes from 2025-03-24 17:07 UTC to 17:13 UTC. Attached is a picture of the unplugged and powered off equipment as I left it.
This morning at 09:04:42 Mon 24mar2025 we had a BSC3 sensor gitch, which VACSTAT normally logs as a single gauge event and does not send alarms. Today however many LVEA gauges also tripped and alarms were sent to the Vacuum Team.
The reason is that on 13mar2025 we had a lock-loss due to an LVEA vacuum event which was below VACSTAT's nominal trip level. To counter this some LVEA slope trip levels were reduced from 1.0e-10 to 3.0e-12 (https://alog.ligo-wa.caltech.edu/aLOG/index.php?callRep=83366)
Fast forward to today and the sequence was:
09:04:42 standard BSC3 sensor glitch, num_glitched = 1, all other gauges set to sensitive mode, meaning 3.0e-13 for most LVEA gauges
09:05:03 2nd LVEA sensor glitched on noise while in sensitive mode, num_glitched = 2, Alarms sent to Vacuum Team
09:11:07 3rd LVEA sensor glitched on noise while in sensitive mode
09:12:37 4th LVEA sensor glitched on noise while in sensitive mode
09:18:35 5th LVEA sensor glitched on noise while in sensitive mode
Vacstat code was changed to permit a sensitivity multiplier to be applied on a gauge-by-gauge basis. The multiplier for those LVEA gauges which have a 3.0e-12 slope trip was set to 1.0 (i.e. no additional sensitivity following a BSC3 trip). The increased sensitivity for the Delta-P trip level was retained.
prod.yaml
---
ifo: H1
glitch_monitor:
lookback_times: [60, 300, 600]
proc_period_secs: 10
vacuum_gauges:
default:
glitch_press_rate: [1.0e-10, 1.0e-10, 1.0e-10]
glitch_delta_press: [1.0e-08, 1.0e-08, 1.0e-08]
valid_value_min: 1.0e-10
valid_value_max: 1.0e-04
sensitivity_press_multiplier: 10.0
sensitivity_deltap_multiplier: 10.0
H0:VAC-LY_X0_PT100B_PRESS_TORR:
description: "Corner Station HAM1"
H0:VAC-LY_Y1_PT120B_PRESS_TORR:
description: "Corner Station BSC2"
glitch_press_rate: [3.0e-12, 3.0e-12, 3.0e-12]
sensitivity_press_multiplier: 1.0
H0:VAC-LX_Y8_PT132_MOD2_PRESS_TORR:
description: "Corner Station BSC3"
H0:VAC-LX_Y0_PT110_MOD1_PRESS_TORR:
description: "Corner Station HAM6"
H0:VAC-LY_Y3_PT114B_PRESS_TORR:
description: "Corner Station CP1"
glitch_press_rate: [3.0e-12, 3.0e-12, 3.0e-12]
sensitivity_press_multiplier: 1.0
H0:VAC-LX_X3_PT134B_PRESS_TORR:
description: "Corner Station CP2"
glitch_press_rate: [3.0e-12, 3.0e-12, 3.0e-12]
sensitivity_press_multiplier: 1.0
H0:VAC-LY_Y4_PT124B_PRESS_TORR:
description: "Corner Station Y-Arm"
glitch_press_rate: [3.0e-12, 3.0e-12, 3.0e-12]
sensitivity_press_multiplier: 1.0
H0:VAC-LX_X4_PT144B_PRESS_TORR:
description: "Corner Station X-Arm"
glitch_press_rate: [3.0e-12, 3.0e-12, 3.0e-12]
sensitivity_press_multiplier: 1.0
H0:VAC-MY_Y1_PT243B_PRESS_TORR:
description: "Mid-Y Y1 Beam Tube"
H0:VAC-MY_Y5_PT246B_PRESS_TORR:
description: "Mid-Y Y2 Beam Tube"
H0:VAC-EY_Y6_PT427_MOD1_PRESS_TORR:
description: "End-Y Y6 BT Ion Pump"
H0:VAC-EY_Y1_PT423B_PRESS_TORR:
description: "End-Y Beam Tube"
H0:VAC-EY_Y3_PT410B_PRESS_TORR:
description: "End-Y BSC10"
H0:VAC-MX_X1_PT343B_PRESS_TORR:
description: "Mid-X X1 Beam Tube"
H0:VAC-MX_X5_PT346B_PRESS_TORR:
description: "Mid-X X2 Beam Tube"
H0:VAC-EX_X6_PT527_MOD1_PRESS_TORR:
description: "End-X X6 BT Ion Pump"
H0:VAC-EX_X1_PT523B_PRESS_TORR:
description: "End-X Beam Tube"
H0:VAC-EX_X3_PT510B_PRESS_TORR:
description: "End-X BSC9"
FAMIS 31078
There seems to have been a slight temperature increase in the laser, diode, and anterooms of just under 1 degF over the past day which perhaps correlates with some changes in laser performance. The NPRO output power increased while outputs of both amplifiers decreased all by small amounts, with some of their pump diodes both increasing and decreasing in output (see laser trends). This change also perhaps made the signal as read by the PMC_TRANS and FSS TPDs a bit noisier. The cooling trends look unchanged over this time period. PMC_REFL has also slightly risen over the past several days, starting before this temperature change, which I am yet unsure of the cause.
TITLE: 03/24 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Observing at 151Mpc
OUTGOING OPERATOR: Ibrahim
CURRENT ENVIRONMENT:
SEI_ENV state: CALM
Wind: 8mph Gusts, 3mph 3min avg
Primary useism: 0.02 μm/s
Secondary useism: 0.27 μm/s
QUICK SUMMARY: Looks like the winds finally died down enough for H1 to lock early this morning and has now been observing for 2 hours. Commissioning time planned to start at 15:30.
IFO is LOCKING and WINDY
I was called when the H1 Notify Guardian reached its 2hr threshold. Lock re-acquisition is taking time understandably due to gusts as high as 50mph (attached pic), and there's nothing that can be done about this until wind speeds decline. Since the wind is forecast to wind down over the next 2/3 hours, I've set guardian to keep trying to lock and reset myself as OWL ops. Hopefully, we can get locked by then.
Call 2: IFO is LOCKING in ENVIRONMENT (WIND)
H1 called again due to wind same reason as before. The wind has decreased but not sufficiently for lock reacquisition. I again just reset the OWL. If I get called a third time, I will stay in IDLE due to WIND.
J. Kissel Oli and I are beginning the process for designing damping loops for the A+ O5 BBSS. We're running through the same process that I've been running through for over a decade designing suspension damping loops, in which I build up a noise budget for the optic displacement in all DOFs using input noises for seismic noise, DAC noise, and OSEM sensor noise filtered through said damping loops, all propagated thru the matlab dynamical model of the suspension. The first step along that journey is revisiting all the input noise sources, and making sure we have good model for those. OSEM noise and DAC noise models have recently been validated and updated when I revisited the HLTS damping loop design (see LHO:65687). However, I haven't worked on damping loops for suspensions suspended from a BSC-ISI since 2013, see G1300537 for the QUAD, G1300561 for the BSFM, and G1300621 for the TMTS. In those, I used the 2015 update to the 2005 requirement curve from T1500122 as the input motion. Now, after a decade worth of commissioning and improvements, I figure it's time to show that work here and use it in modeling future SUS damping loops where the SUS is mounted from a BSC-ISI. One of the biggest things we've learned over the decades is that the seismic noise input to the suspension at its "Suspension Point" motion for a given suspension can be the (quadrature) sum of many of the ISI's cartesian degrees of freedom, and depends on where and in what orientation it is on the optical table (see T1100617). As such, we installed front-end infrastructure to calculate the calibrate the lowest stage sensors -- the GS13 inertial sensors -- into both the Cartesian and Euler basis (see E1600028). In this aLOG, I do as I did for the HAM-ISI LHO:65639, I show the Cartesian contributions to each of the Beam Splitter's SUS point motion, by multiplying the Cartesian channels by the coefficients in the CART2EUL matrix for the beam splitter. The time I used for this performance of the H1 ISI BS was 0.01 Hz binwidth (128 sec FFT), 10 average, 50% overlap data set starting at 2025-03-19 14:00 UTC. - This was a late night local set, with no wind and 0.1 [um_BLRMS] level microseism (between 0.1-0.3 Hz) - GND to ST1 Sensor correction is ON, including the DIFF and COMM inputs. - Here at H1, the corner station does NOT have beam rotation sensors to improve the GND T240 sensor correction signal. But, both end stations have a BRS. - The wind was low at this measurement time, but it's worth saying that each end-stations wind fences are in dis-repair at the moment, too be fixed soon. - ST1 Z drive to ST1 RZ T240 decoupling is ON with a "pele_rz" filter - Off diagonal ST1 dispalign matrices are in play, X to RX & RY = -1e-4 & 1e-4, Y to RX = -7e-4, Z to RX & RY = 3.5e-3 & 2.5e-3 - ST1 Blend Filters: - X & Y = nol4cQuite_250 - Z = 45mHz_cps - RX & RY = Quite_250_cps - RZ = nol4cQuite_250. - As far as I can tell, there's NO ST1 to ST2 sensor correction on the ST2 CPS, nor is there and ST1 to ST2 FF to the ST2 actuators. - ST2 Blend Filters: - X & Y = 250mhz - Z = 250mhz - RX & RY = tilt_800b - RZ = 250mhz These will be used to make updates to /ligo/svncommon/SusSVN/sus/trunk/Common/MatlabTools/ seisBSC.m or seisBSC2.m which are toy models of the BSC-ISI performance, used so you don't have to carry around some giant .mat file of performance and you can per-interpolate on to an arbitrary frequency vector, much like I did for seisHAM.m in CSWG:11236. I've committed the .xmls and .pngs in the following SeiSVN directory: /ligo/svncommon/SeiSVN/seismic/BSC-ISI/H1/BS/Data/Spectra/Isolated/ASD_20250319/
Dear Oli,
It may be useful to remember that when Jeff says that the "input to the suspension at its "Suspension Point" motion for a given suspension can be the (quadrature) sum of many of the ISI's cartesian degrees of freedom" - what he means is that, if you want to make a Statistical model (which you do), and if the DOFs are independant (which maybe they are, and maybe they are not), then using the quadruture sum of the ASDs is a reasonable thing to do. In fact, the SUSpoint in reality, and the calculation of the SUSpoint, are done with a linear combination, NOT a quadrature sum. This means that if you grab some data from the cart basis sensors, take the ASDs (where you lose the phase), and add them in quadrature you will NOT get the ASD of the measured suspoint. I think this difference is not going to impact any of your calculations, but maybe it will help you avoid aggravation if you try to do some double checking.
-Brian
The Cartesian performance ASDs of the ISI BS to be used in the statistical model (in the way that Brian cautions in LHO:83473 above) have been exported to /ligo/svncommon/SeiSVN/seismic/BSC-ISI/H1/BS/Data/Spectra/Isolated/ASD_20250319/ 2025-03-19_1400UTC_H1SUSBS_CART_XYZRXRYRZ_ASD.txt (in the DOF order mentioned in the filename.) In the same directory, I also export the ASD of live, projected, coherent linear sum computed by the front-end 2025-03-19_1400UTC_H1SUSBS_EUL_LTVRPY_ASD.txt (in the DOF order mentioned in the filename.) If someone wants to race me, they can use this data and the CART2EUL matrix from the screenshot in LHO:83470, or if you want it programmatically, use /opt/rtcds/userapps/release/isc/common/projections/ ISI2SUS_projection_file.mat and running the following in the matlab command line, >> load /opt/rtcds/userapps/release/isc/common/projections/ISI2SUS_projection_file.mat >> ISI2SUSprojections.h1.bs.CART2EUL ans = -0.7071 0.7071 -0.2738 0 0.1572 0.1572 -0.7071 -0.7071 -0.0173 0 -0.1572 0.1572 0 0 0 1 -0.2058 0.1814 0 0 0 0 -0.7071 0.7071 0 0 0 0 -0.7071 -0.7071 0 0 1 0 0 0 ... but if I win the race, this plot will be a good by-product of the updates to seisBSC.m, which I'll likely post to the CSWG aLOG, like I did for seisHAM.m in CSWG:11236.
Jim reminds me of the following: - This BSC-ISI, ISIB2 has been performing poorly since ~2020. For some yet-to-be-identified reason, after years of physical, electronic, and data analysis investigations by Jim -- see IIET:15234 -- his best guess is some sort of mechanical "rubbing," i.e. mechanical interference / shorting of the seismic isolation, typically by cables. - He points is finger at the H2 corner (use T1000388 to reminder yourself of where that is on BSC2). - You can use the "Network" summary pages (https://ldas-jobs.ligo.caltech.edu/~detchar/summary/) and navigate to "Today" > "SEI" tab > "Summary [X]" or "Summary [Y]" or "Summary [Z]" pages, and look at the bottom row of plots to see how the ISIBS compares against other ISIs at LHO (left plot) and LLO (right plot). Here's a direct link to the plots including 2025-03-19 at 14:00 UTC, with the with the "SEI Quiet" time restriction mode ON. - Also, remember that the MICH lock-acquisition drive from the M2 OSEMs on the SUSBS causes back-reaction on the cage, which messes with the ISI controls, the ISIBS's isolation state guardian is regularly in the FULLY_ISOLATED_SO_ST2_BOOST state, which leaves the FM8 "Boost_3" off until after the ISC_LOCK guardian requests SEI_BS to FULLY_ISOLATED. Because I took data during nominal low noise, the ISI was fully isolated. However, the summary pages above -- even in SEI Quiet mode -- don't filter for whether the ISI is in FULLY_ISOLATED, so you'll that the ISIBS is consistently performing worse. *This* is not a fair comparison or show of how the ISIBS performs worse that the other BSC-ISIs, so take the plots with a big grain of salt. Also, another point of configuration notes: - This ISI, like all ISIs at LHO have their CPS synchronized to the timing system.