Displaying reports 1-20 of 82873.Go to page 1 2 3 4 5 6 7 8 9 10 End
Reports until 08:53, Wednesday 25 June 2025
H1 SQZ
oli.patane@LIGO.ORG - posted 08:53, Wednesday 25 June 2025 (85320)
No SQZ Range Comparison

I just took some No SQZing time (starting at 2025-06-25 15:01 UTC (1427040078)) and compared it to the pre-vent No SQZ Time of 2025-03-26 16:01 UTC (1427040078). I used the command python3 range_compare.py 1427040078 1434898893 --span 600. I originally wanted to use a time span of half an hour instead of 10 minutes, but the March time had some sort of glitch that messed up its range calculation, and the range calculation for today's time was the same, so I went with 10 minutes. Here's the range comparison.

Non-image files attached to this report
H1 General
oli.patane@LIGO.ORG - posted 07:34, Wednesday 25 June 2025 - last comment - 08:24, Wednesday 25 June 2025(85318)
Ops Day Shift Start

TITLE: 06/25 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Observing at 145Mpc
OUTGOING OPERATOR: TJ
CURRENT ENVIRONMENT:
    SEI_ENV state: CALM
    Wind: 5mph Gusts, 2mph 3min avg
    Primary useism: 0.01 μm/s
    Secondary useism: 0.06 μm/s
QUICK SUMMARY:

Observing and have been Locked for almost 2.5 hours. Range is okay at 145Mpc, it looks like its been drifting a bit down since the lock started.

Comments related to this report
oli.patane@LIGO.ORG - 08:24, Wednesday 25 June 2025 (85319)ISC, Lockloss

Looking at the lockloss from last night (2025-06-25 11:02 UTC - note that the 'refined lockloss time' is one whole second early than the actual lockloss time), it's immediately clear that we lost lock due to a quick ringup, but it is unclear where that ringup came from or what frequency it was actually at. We can see an 8Hz oscillation in DARM one second before the lockloss, but looking at the LSC channels, SRCL sees a 1.5 Hz oscillation right before the lockloss, and PRCL has a 4.5 Hz oscillation (although the PRCL oscillation could be unrelated, although it does look like it grows a bit larger). In the ASC channels, MICH P OUT has a ringup at ~3.8 Hz in that last second, and some of the other ASC channels look like they maybe also have some sort of excursion right before the lockloss.

Images attached to this comment
LHO VE (VE)
gerardo.moreno@LIGO.ORG - posted 02:14, Wednesday 25 June 2025 (85317)
Annulus Ion Pump for BSC1 Rails

Signal railed about 5:18 PM local time, I checked trend data for PT120 and PT180 and no pressure rise noted inside the main volume.  Attached is 3 day trend of the pump behavior, very glitchy for a long while already.

System will be evaluated as soon as possible.  AIP last replaced on 2015, see aLOG 18261.

Images attached to this report
H1 General (Lockloss)
anthony.sanchez@LIGO.ORG - posted 21:23, Tuesday 24 June 2025 (85315)
Lockloss from NLN

Lockloss From Nominal_Low_Noise @ 03:01:08 UTC.

I'm not exactly sure what caused this LL .

Images attached to this report
LHO VE (VE)
travis.sadecki@LIGO.ORG - posted 20:35, Tuesday 24 June 2025 (85314)
HAM1 turbo off, pump cart removed, and PT-100 gauge pigtail installed

With the HAM1 ion pump performing nobly, the HAM1 turbo pump was spooled down and the associated SS-500 pump cart was disconnected from the HAM1 volume.  All mechanical sources of vacuum equipment noise should be in nominal observing mode.  

The HAM1 pressure gauge, PT-100, was connected via a standard gauge pigtail cable, directly from gauge body sub-D to supply CPC.  All of the temporary cabling was removed including the breakout board and analog voltage readout pigtail, so PT-100 is solely on Ethercat communication.  

H1 SUS (SUS)
anthony.sanchez@LIGO.ORG - posted 17:59, Tuesday 24 June 2025 (85311)
Weekly In-Lock SUS Charge Measurement

Famis 28411 Weekly In-Lock SUS Charge Measurement.

This command is very useful to determine if the SUS charge measurements ran in the last week:
ls /opt/rtcds/userapps/release/sus/common/scripts/quad/InLockChargeMeasurements/rec_LHO  -lt | head -n 6

 

Coherence for bias_drive_bias_off is 0.06426009151210332, which is below the threshold of 0.1. Skipping this measurement
Cannot calculate beta/beta2 because some measurements failed or have insufficient coherence!
Cannot calculate alpha/gamma because some measurements failed or have insufficient coherence!
Something went wrong with analysis, skipping ITMX_13_Hz_1434811847

Images attached to this report
H1 General
oli.patane@LIGO.ORG - posted 17:41, Tuesday 24 June 2025 (85310)
Ops Day Shift End

TITLE: 06/25 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Observing at 145Mpc
INCOMING OPERATOR: Tony
SHIFT SUMMARY:

Currently Observing at 150Mpc and have been Locked for just over 2 hours. Relocking went pretty smooth, with the only issue being DRMI/PRMI taking very long to catch still. During maintenance today, the auxiliary pumps around HAM1 were turned off, and the jitter noise that we had been seeing around 550-600Hz is basically almost fully gone.


LOG:

14:30 Observing at 145Mpc and have been Locked for almost 4 hours
14:45 Out of Observing to run SUS Charge measurements
15:00 Lockloss during SWAP_BACK_ETMX
    - It doesnt look like it was due to anything that was happening with the swap

19:31 Started relocking
    - Initial alignment
    - Lockloss from MOVE_SPOTS
22:33 NOMINAL_LOW_NOISE
22:34 Observing

Start Time System Name Location Lazer_Haz Task Time End
15:03 FAC Chris, Eric XARM n Bee-am sealing 17:32
15:09 FAC Randy LVEA n Drilling holes in a pipe 16:17
15:10 FAC Nellie, Kim LVEA n Tech clean 16:27
15:21 VAC Gerardo, Jordan LVEA, FCETube n Turning off HAM1 pump, replacing gauge in FCETube, craning onto termination slab 19:25
15:33 VAC Janos EX, EY n Pumping line parts collection 18:40
15:39 FAC Tyler XARM n Driving Big Red to move spool 16:39
15:45 SQZ Camilla, Sheila LVEA y(local) SQZT7 work 18:49
15:45 PSL Jason LVEA n Inventory 17:32
15:46 FAC Mitchell LVEA n Inventory 16:09
15:53 CDS Jonathan remote n Rebuilding camera software 16:01
15:54 VAC Travis LVEA n Helping Gerardo and Jordan 17:21
16:07 SUS RyanC CR N OPLEV charge measurements, EY & EX 17:38
16:14 FAC Richard, Ken LVEA n Looking at cable tray near HAM6 16:27
16:15   Christina LVEA n Inventory 16:33
16:16 VAC Jackie FCETube n Joining Jordan and Gerardo 17:14
16:18 CDS Patrick remote n Reimaging VAC Beckhoff computer 19:05
16:33 FAC Randy MX n Craning 18:20
16:33   Christina MY n Taking a photo 17:01
16:36 SEI Jim remote n HAM1 ISI measurements 19:05
16:52 FAC Mitchell FCETube n Inventory 17:59
16:52 FAC Tyler XARM n Meeting up with Chris, Eric, and the bees 17:07
16:57   Keita LVEA n Checking analog camera inventory 17:33
16:59 FAC Kim, Nellie HAM Shack n Tech clean 17:27
17:05 EE Fil, Marc LVEA n Fixing BSC temp sensor 18:35
17:11   Jennie, Leo, Tooba LVEA n Tour 17:42
17:24   Betsy, Brian O, LIGO India people LVEA n LIGO India planning 18:09
17:24   Christina LVEA n Inventory 17:33
17:27 FAC Kim MX n Tech clean 18:28
17:28 FAC Nellie MY n Tech clean 18:58
17:34   Christina HAM Shack n Looking for a forklift 18:17
17:42 SQZ Jennie, Leo LVEA y(local) Joining on SQZT7 18:31
18:16   Betsy + others EY n Looking at stuff 19:05
18:18   Christina LVEA n Still looking for a forklift 18:48
18:49   Camilla OpticsLab n   18:51
19:11   RyanC LVEA n Sweep 19:35
19:37   Matt LVEA n Unplugging a cable 19:40
19:48 EE Fil MX n Dropping stuff off 20:18
20:19 FAC Chris MY, EY n Checking filters 21:16
20:23 FAC Tyler XARM n Checking out more bees 20:51
20:36 VAC Gerardo, Jordan FCETube n Opening FC gate valve 21:09
22:10 PCAL Tony PCAL Lab y(local) Taking stuff 22:16
H1 GRD (ISC)
elenna.capote@LIGO.ORG - posted 17:38, Tuesday 24 June 2025 (85309)
Guardian edits to speed up ADS

Ryan S., Elenna

Ryan and I are still trying to speed up the MOVE_SPOTS state. Today, Ryan implemented new code that checks the convergence of the loops and only ramps up the ADS gains of loops that are not yet converged to help them converge faster. This appeared to work well, although the state is still slow. We are now taking the spots to the FINAL spots that the camera servos go to, instead of some old spot, so it's possible that which loops that are far off have changed.

Ryan also pointed out that the ENGAGE_ASC_FOR_FULL_IFO state is taking a while because it is limited by the convergence of the PIT3 ADS. This is likely because the POP A offset used in DRMI ASC is not quite right, so I adjusted it for pitch so the PRM should be closer to the full lock position. SDFed.

Images attached to this report
H1 SUS
oli.patane@LIGO.ORG - posted 17:30, Tuesday 24 June 2025 (85288)
SR3 SUSPOINT TFs and regular damping TFs taken for estimator work

The last work that was done for the estimator was back on May 21st 84548. The data that I got then was of the Open Loop TFs for when SR3 OSEMINF gains were nominal (aka what they are now) vs our 'more calibrated' values that are shown in 84367. I then wrote a script that takes in the OLTF traces that were exported from diaggui and divides them for each DOF (results). The average between the two divided traces is what we are going to be using as the gain values needed to counter the 'more calibrated' values. I put these gains in as FM7(previously had old unused gains in) in the DAMP filter bank.

Today I changed the M1 coil driver state to 1, set the OSEMINF gains to the ones listed in 84367,  turned on FM7 for all degrees of freedom in the DAMP filter bank to turn on the gain corrections, changed the gain for DAMP Y to be -0.1 instead of the nominal -0.5, and then took SUSPOINT to M1 transfer function measurements.The measurements can be found in: /ligo/svncommon/SusSVN/sus/trunk/HLTS/H1/SR3/Common/Data/2025-06-24_1700_H1ISIHAM5_ST1_WhiteNoise_SR3SusPoint_{L,T,V,R,P,Y}_0p02to50Hz.xml, in svn as r12356.

After that, I took regular transfer functions with damping on for SR3 M1 to M1. Unfortunately I accidentally took the transfer functions for Pitch with the wrong bin width, so the original xml file there has double the values as the other ones. The saved .mat file, though, has the data all the same size. The measurements can be found in: /ligo/svncommon/SusSVN/sus/trunk/HLTS/H1/SR3/SAGM1/Data/2025-06-24_1900_H1SUSSR3_M1_WhiteNoise_{L,T,V,R,P,Y}_0p02to50Hz.xml, in svn as r12359. The results are found in: /ligo/svncommon/SusSVN/sus/trunk/HLTS/H1/SR3/SAGM1/Results/2025-06-24_1900_H1SUSSR3_M1_ALL_TFs.pdf, also r12359.

After these measurements, everything was put back to how it was before, including the OSEMINF gains, the DAMP Y gain, and the DAMP FM7 filters turned back off.

 

Images attached to this report
H1 ISC
anthony.sanchez@LIGO.ORG - posted 17:09, Tuesday 24 June 2025 - last comment - 09:21, Wednesday 25 June 2025(85307)
Range Comparison: April 2024 vs Today.

Range comparison plot:
Command Ran:  python3 range_compare.py 1396833438 1434842418 --span 1800
time 1 is April 11th 2024
time 2 is today at 23:20 UTC June 24th 2025
 

Non-image files attached to this report
Comments related to this report
elenna.capote@LIGO.ORG - 17:25, Tuesday 24 June 2025 (85308)

I ran two brucos, one on NOLINES and one on CLEAN.

bruco NOLINES

bruco CLEAN

Overall, the message of Tony's alog is that, relative to our best range, we have lost 10 Mpc by the time we reach 100 Hz, and an additional 5 Mpc by the time we reach 1 kHz. The brucos above show a lot of low level coherence with: MICH, SRCL, PRCL, REFL RIN. There is a chance that making additional improvements to the feedforward can help. Right now it's hard to tell how many Mpc that gets us back, but it's where we should start.

oli.patane@LIGO.ORG - 09:21, Wednesday 25 June 2025 (85321)

I made a range comparison with a time form last night when we got up to 155Mpc (Jun 25, 2025 05:07:30 UTC (1434863268)) and compared it to a time of good range before the vent, Apr 01, 2025 03:05:42 UTC (1427511960). Here's the range comparison.

Non-image files attached to this comment
LHO VE (VE)
gerardo.moreno@LIGO.ORG - posted 16:45, Tuesday 24 June 2025 - last comment - 16:50, Tuesday 24 June 2025(85303)
Filter Cavity Tube Bad Gauge Removal and Replacement-Replacement

(Jordan V., Gerardo M.)

Today we replaced the MKS gauge at FC-C-1, this is the first 6 way cross inside the filter cavity tube enclosure, we installed serial number 390F00490, twice, yes two times.  It turns out that the flange has some scratches on the knife edge, and it was not going to seal regardless of the effort that we put into it.  Once the gauge was removed the scratches transferred to the copper gasket.  We replaced it with serial number 390F00495, this one seems to be doing good.  New conflat was leak tested and no leak was detectable above 2.42e-10 torr*l/sec.

The old gauge serial number is 390F00406 with a date code of June 2021.

Images attached to this report
Comments related to this report
jordan.vanosky@LIGO.ORG - 16:50, Tuesday 24 June 2025 (85305)

Additional pictures of the knife edge damage/dirty flange from manufacturer.

Images attached to this comment
H1 General
anthony.sanchez@LIGO.ORG - posted 16:38, Tuesday 24 June 2025 (85302)
Tuesday Ops Eve Shift.


TITLE: 06/24 Day Shift: 1430-2330 UTC (0730-1630 PST), all times posted in UTC
STATE of H1: Observing at 148Mpc
OUTGOING OPERATOR: Oli
CURRENT ENVIRONMENT:
    SEI_ENV state: CALM
    Wind: 11mph Gusts, 7mph 3min avg
    Primary useism: 0.02 μm/s
    Secondary useism: 0.06 μm/s
QUICK SUMMARY:
H1 has been Locked for close to an hour.

Note:
CDS Alarm H0:VAC-LY_Y3_PT114B_PRESS_TOUR  alarmed at 1434843240.  Oli metioned that Davis said this is nothing to worry about, I'm just documenting it here.
Everything else looks fine.
 

 

H1 SQZ
camilla.compton@LIGO.ORG - posted 15:31, Tuesday 24 June 2025 - last comment - 19:51, Tuesday 24 June 2025(85295)
7dB on SQZ HD with New Spot in OPO Crystal

Sheila, Camilla. WP#12573

Sheila and I realigned the Homodyne after it needed to be touched to re-seat one of the PDs in 85116. Sheila balanced powers using the LO waveplate and measured visibility to be 98.4% (11.4mV min 1.46 V max).

We then used SQZ_MANAGER SQZ_READY_HD (needed to change LO gain, see sdf). Attached plot.

Last done by Kevin/Vicky in 84661, they got 6.4 to 7 dB of SQZ, 14.6 dB of anti-SQZ, and 12.1 dB of mean squeezing.

Type NLG Angle SQZ (@300Hz) DTT Ref
LO shot noise N/A N/A Used as 0dB ref 1
ASQZ 10 (+) 204 14.3dB ref 2
SQZ 10 (-) 114 -7dB ref 3
Mean SQZ 10 N/A 10.7dB ref 4
 
OPO Setpoint Amplified Max Amplified Min UnAmp Dark NLG OPO Gain
95uW 0.0530 0.001913 0.005307 9.1e-5 10.16 -8
Images attached to this report
Comments related to this report
kevin.kuns@LIGO.ORG - 19:51, Tuesday 24 June 2025 (85313)

After Sheila and Camilla realigned the homodyne, I tried estimating the NLG by looking at the ADF beatnote while varying the squeezing angle to map out the ADF-LO ellipse as described in the ADF paper.

The channels used to measure the beatnote were H1:SQZ-ADF_HD_DIFF_NORM_{I,Q}. The squeeze angle was varied by sweeping H1:SQZ-CLF_REFL_RF6_PHASE_PHASEDEG from 0 to 275 degrees with both polarities of the CLF servo (H1:SQZ-CLF_SERVO_IN2POL).

I then fit the data to an ellipse. The ratio of the semi-major to semi-minor axes is denoted by G in the ADF paper and was then used to compute the NLG. This results in
G: 5
NLG: 8.8

By comparison, an NLG of 10.2 is equivalent to G=5.4, which is consitent with the accuracy of this data. This kind of measurement could probably be improved in the future by taking more data near the anti-squeezing points on the ellipse.

It's not obvious from the ADF paper how to convert from G to NLG and back. For posterity, the functions I used for these conversions are also attached. These use the "alternative OPO configuration" shown in Fig 6(a) of the ADF paper.

The data is located at /ligo/gitcommon/squeezing/sqzutils/data/NLG_HD_06_24_2025.h5.

Images attached to this comment
Non-image files attached to this comment
H1 CDS
david.barker@LIGO.ORG - posted 14:41, Tuesday 24 June 2025 - last comment - 16:42, Tuesday 24 June 2025(85296)
CDS Maintenance Summary: Tuesday 24th June 2025

WP12623 h1asc add fast channels to DAQ

Elenna, Dave:

A new h1asc model was rev-locked and installed. Four new fast DQ channels were added to the DAQ (channel, rate):

> H1:ASC-DC6_P_IN1_DQ, 256
> H1:ASC-DC6_P_OUT_DQ, 512
> H1:ASC-DC6_Y_IN1_DQ, 256
> H1:ASC-DC6_Y_OUT_DQ, 512

DAQ restart needed.

WP12570 Restart Digivideo Cameras with latest pylon

Patrick, Jonathan, Dave:

Jonathan updated pylon on h1digivideo[4,5,6] and restarted all the camera servers on these machines. This should fix the bug of stuck open files accumulating when the camera connection is interrupted.

No DAQ restart needed.

Add PID SMOO channels to vacuum SDF

Dave:

Prior to today's h0vacly restart I added the missing CP PID-control SMOO channels to the vacuum SDF monitor.req and safe.snap files. SDF was restarted 08:29. No DAQ restart needed.

WP12577, 12608, 12615 Upgrade LY Vacuum Controls

Janos, Gerardo, Patrick, Jonathan, Erik, Dave:

Patrick installed a new h0vacly system this morning. Main items are:

Pleae see Patrick's alog for details.

A extended DAQ restart was required, renaming Ion Pump raw minute trend files for uninterrupted lookback and construcing new PT100 (HAM1) raw minute trends following the upgrade of h0vaclx last Tuesday (17th June 2025).

DAQ Restart

Jonathan, Erik, Patrick, Dave:

Immediately following the restart of h1asc at 11:52, the DAQ was restarted using the following procedure:

  1. both trend writers were turned off, allowing raw minute trend file changes on a static system
  2. ip_copy_scripts.bsh was ran on both TWs, for each Ion Pump channel which is being renamed, this copied the old file into the correct directory for the new file, preserving recent lookbacks.
  3. Erik ran his script to merge the previous HAM1 PT100 data sets into the new channel name.
  4. A new H0EPICS_VACLY.ini was generated
  5. 0-leg was restarted, followed by a EDC restart on h1susauxb123
  6. 1-leg was restarted, requiring a second restart of GDS1.

It was at this late point that I remembered that the temporary H1 version of PT100B is no longer needed, and indeed this channel has no data following the removal of the PT100B Volts channel from h0vacly. However since it is still in the EDC, we need to continue running the temporary IOC until the next DAQ restart. I've removed it from edcumaster.txt as a reminder.

GPS Leap Seconds Updates

Jonathan, Erik, Dave:

Erik's FAMIS task reminded us that the leapseonds files expiration date of 30 June 2025 is rapidly approaching. Although no leap seconds are to be applied, the files need to be updated to reset their expiration dates. Please see Erik and Jonathan's alog for more details.

DNS testing 

Erik:

ns1 (backup DNS server) was used by Erik to see if we can reproduce the error whereby loss of connection to GC caused internal CDS name resolution issues. It did not.

Comments related to this report
david.barker@LIGO.ORG - 16:42, Tuesday 24 June 2025 (85304)

Vacuum Ion Pump channel name changes (old-name, new-name)

 

H0:VAC-FCES_IP23_II123_AIP_IC_VOLTS H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_VOLTS
H0:VAC-FCES_IP23_II123_AIP_IC_VOLTS_ERROR H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_VOLTS_ERROR
H0:VAC-FCES_IP23_II123_AIP_IC_MA H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_MA
H0:VAC-FCES_IP23_II123_AIP_IC_MA_ERROR H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_MA_ERROR
H0:VAC-FCES_IP23_II123_AIP_IC_LOGMA H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_LOGMA
H0:VAC-FCES_IP23_II123_AIP_IC_LOGMA_ERROR H0:VAC-FCES_IPFCC9_IIC9_AIP_IC_LOGMA_ERROR
H0:VAC-FCES_IP23_VI123_AIP_PRESS_TORR H0:VAC-FCES_IPFCC9_VIC9_AIP_PRESS_TORR
H0:VAC-FCES_IP23_VI123_AIP_PRESS_TORR_ERROR H0:VAC-FCES_IPFCC9_VIC9_AIP_PRESS_TORR_ERROR
H0:VAC-FCES_IP24_II124_AIP_IC_VOLTS H0:VAC-FCES_IPFCD1_IID1_AIP_IC_VOLTS
H0:VAC-FCES_IP24_II124_AIP_IC_VOLTS_ERROR H0:VAC-FCES_IPFCD1_IID1_AIP_IC_VOLTS_ERROR
H0:VAC-FCES_IP24_II124_AIP_IC_MA H0:VAC-FCES_IPFCD1_IID1_AIP_IC_MA
H0:VAC-FCES_IP24_II124_AIP_IC_MA_ERROR H0:VAC-FCES_IPFCD1_IID1_AIP_IC_MA_ERROR
H0:VAC-FCES_IP24_II124_AIP_IC_LOGMA H0:VAC-FCES_IPFCD1_IID1_AIP_IC_LOGMA
H0:VAC-FCES_IP24_II124_AIP_IC_LOGMA_ERROR H0:VAC-FCES_IPFCD1_IID1_AIP_IC_LOGMA_ERROR
H0:VAC-FCES_IP24_VI124_AIP_PRESS_TORR H0:VAC-FCES_IPFCD1_VID1_AIP_PRESS_TORR
H0:VAC-FCES_IP24_VI124_AIP_PRESS_TORR_ERROR H0:VAC-FCES_IPFCD1_VID1_AIP_PRESS_TORR_ERROR
H0:VAC-FCES_IP25_CS187_STATUS H0:VAC-FCES_IPFCH8A_CSH8A_STATUS
H0:VAC-FCES_IP25_II187_IC_VOLTS H0:VAC-FCES_IPFCH8A_IIH8A_IC_VOLTS
H0:VAC-FCES_IP25_II187_IC_VOLTS_ERROR H0:VAC-FCES_IPFCH8A_IIH8A_IC_VOLTS_ERROR
H0:VAC-FCES_IP25_II187_IC_AMPS H0:VAC-FCES_IPFCH8A_IIH8A_IC_AMPS
H0:VAC-FCES_IP25_II187_IC_AMPS_ERROR H0:VAC-FCES_IPFCH8A_IIH8A_IC_AMPS_ERROR
H0:VAC-FCES_IP25_VI187_PRESS_TORR H0:VAC-FCES_IPFCH8A_VIH8A_PRESS_TORR
H0:VAC-FCES_IP25_VI187_PRESS_TORR_ERROR H0:VAC-FCES_IPFCH8A_VIH8A_PRESS_TORR_ERROR

 

H1 TCS (DetChar-Request)
camilla.compton@LIGO.ORG - posted 13:20, Monday 23 June 2025 - last comment - 18:30, Tuesday 24 June 2025(85247)
2Hz Comb in DARM from ITMX HWS Camera?

Ansel, Sheila, Camilla

Last week, Ansel noticed that there is a 2Hz comb in DARM since the break, similar to that that we've seen from the HWS camera sync frequency and power supplies and fixed in 75876.  The cabling has not been changed since, the camera sync frequency has been changed.

Our current camera sync frequencies are: ITMX = 2Hz, ITMY = 10Hz. We have typically seen these combs in H1:PEM-CS_MAG_LVEA_OUTPUTOPTICS_Y_DQ. With a 0.0005Hz BW on DTT I can't easily see these combs, see attached.

Images attached to this report
Comments related to this report
evan.goetz@LIGO.ORG - 13:31, Monday 23 June 2025 (85254)DetChar
It may be difficult to see in a standard spectrum, but can be clearly seen in Fscan plots linked off of the summary pages. For the "observing" Fscan, the interactive spectrum plot shows the 2 Hz comb marked automatically. See the attached image of H1:GDS-CALIB_STRAIN_CLEAN
Images attached to this comment
camilla.compton@LIGO.ORG - 17:04, Tuesday 24 June 2025 (85306)

Verifed that the cabling has not changed since 75876.

Next steps we should follow, as listed in 75876 would be to try using a different power supply or lowering the voltage to +12V. Or, there is a note suggesting Fil could make a new cable to power both the camera and CLink's via the external supply (14V is fine for both). 

evan.goetz@LIGO.ORG - 18:30, Tuesday 24 June 2025 (85312)DetChar
Thanks Camilla. If anything can be done more rapidly than waiting another week, it would be very much appreciated. Continuing to collect contaminated data is bad for CW searches.
Displaying reports 1-20 of 82873.Go to page 1 2 3 4 5 6 7 8 9 10 End