Displaying reports 51221-51240 of 86130.Go to page Start 2558 2559 2560 2561 2562 2563 2564 2565 2566 End
Reports until 22:04, Sunday 30 April 2017
H1 DCS (DCS)
gregory.mendell@LIGO.ORG - posted 22:04, Sunday 30 April 2017 - last comment - 22:11, Sunday 30 April 2017(35924)
Problem with DMT H1 lock segments and redundant hoft at LHO
FYI:

1. Note that the good news is I see H1 low latency hoft making it to CIT, and it is being aggregated. Also it seems the low-latency analysis can determine the lock segments from the hoft frames. 

Thus, the problems listed below are something we can investigate Monday morning. An email has been sent to DASWG and the relevant persons about this.

2. However, the redundant copy of hoft written by the DMT seems to have
stopped today at:

Apr 30 11:21 PDT.

Much earlier today, the DMT logs report Input/output error like this:

$ tail  logs/H1_rename_Frames-2017.04.25-09.43.log

‘/gds-h1/dmt/frames/hoft/H1/.temp/H-H1_DMT_C00-1177574268-4.gwf’ ->
‘/gds-h1/dmt/frames/hoft/H1//H-H1_DMT_C00-117757/H-H1_DMT_C00-1177574268-4.gwf’
mv: cannot move
‘/gds-h1/dmt/frames/hoft/H1/.temp/H-H1_DMT_C00-1177574268-4.gwf’ to
‘/gds-h1/dmt/frames/hoft/H1//H-H1_DMT_C00-117757/H-H1_DMT_C00-1177574268-4.gwf’:
Input/output error

3. Also, thus summary pages show dropout of the H1 lock segments

https://ldas-jobs.ligo-wa.caltech.edu/~detchar/summary/day/20170430/

and nagios reports errors like,

https://dashboard.ligo.org/cgi-bin/nagios3/status.cgi?host=all&servicestatustypes=28

"Known segments in frames that aren't in database"

Thus, low-latency data is making it to CIT, I think with the correct
locked segments, but these are not making into the segment database.

This is something we'll investigate Monday morning.

 

Comments related to this report
gregory.mendell@LIGO.ORG - 22:11, Sunday 30 April 2017 (35925)

This is not related to previous problems with the calibration code not reporting lock after a data dropout.

This recent frame,

bash-4.4$ ls -l /scratch/frames/cache/hoft/H1/H-H1_DMT_C00-117764/H-H1_DMT_C00-1177648628-4.gwf
-rw-r--r--   1 40829    40829     392952 Apr 30 21:37 /scratch/frames/cache/hoft/H1/H-H1_DMT_C00-117764/H-H1_DMT_C00-1177648628-4.gwf

bash-4.4$ tconvert -l 1177648628
Apr 30 2017 21:36:52 PDT

reports a H1:GDS-CALIB_STATE_VECTOR of

67010559 = 0b11111111100111111111111111

which says H1 is in analysis ready mode.

H1 General
nutsinee.kijbunchoo@LIGO.ORG - posted 21:01, Sunday 30 April 2017 (35923)
Ops Eve Midshift Summary

Been locked and Observe for almost 52 hours. The wind is still~30 mph.

H1 ISC
terra.hardwick@LIGO.ORG - posted 18:20, Sunday 30 April 2017 (35837)
PI Mode23 coupling

Amplitude of 32761 Hz PI (Mode23 in LHO damping scheme) correlates with 3 - 30 Hz seismic motion. There is also a corresponding comb that rises in the PI OMC channel at 2*(Nyquist - PI frequency), where Nyquist = 32768 Hz and PI freq ~ 32761 Hz. 

Thanks to the summary pages, I'd noticed that the 28 - 32 kHz PI band somewhat correlated with SenseMon range. I narrowed this correlation down to 3 - 30 Hz seismic motion (which we now know correlates with range due to input beam scattering) and PI ETMY Mode 23 aka ~32761 Hz. Correlation seems to be only with this mode, ruling out electronics coupling since a mode sensed right next door in frequency world has no hardware difference - both are sensed with H1:OMC-PI_DCPD_64KHZ_AHF_DQ. First attachment is 32 hr stretch showing correlation between narrow OMC-PI band, seismic, and range. 

Kiwamu then did a drive/damp test of this mode and found narrow lines appeared in DARM while driving/damping well below saturation. Looking at the OMC channel, he actually saw just the first peak of a 14.5 Hz comb seen in OMC-PI, as well as 64, 78.5, 142.5 Hz lines. Second attachment shows comb while he was driving. This ~15 Hz comb is visible in OMC-PI anytime Mode23 is rung up enough (haven't quantified 'enough' yet) and first peak is always 2x the frequency difference of 32768 Hz (OMC-PI's Nyquist) and Mode23 frequency; I've confirmed this to within .01 Hz.  Third attachment shows presence of first peak in comb during a typical day's Mode23 non-driven ring up. I haven't seen the other peaks Kiwamu saw (64, etc.) any other time.

This is an actual mode ringing up (as opposed to just changing in sensing at OMC) as it's seen in the TransMon QPD's (and all feed forward is off for this mode). These have much lower SNR for PI so mode has to be unusually high to see. Fourth attachment shows a recent day with high 3 - 30 Hz seismic activity where amplitude correlations are visible everywhere. Fifth attachment shows spectrum during the peak just after 5 hours in previous attachment. Note that in TransMon channel, signal has been demoded with 32800 Hz, so two peaks are (32800 - 32761) = 39 Hz and (32800 + 32761 - 65536) = 25 Hz.

Images attached to this report
H1 General
edmond.merilh@LIGO.ORG - posted 16:22, Sunday 30 April 2017 (35921)
Shift Summary - Day

TITLE: 04/30 Day Shift: 15:00-23:00 UTC (08:00-16:00 PST), all times posted in UTC
STATE of H1: Observing at 51Mpc
INCOMING OPERATOR: Nutsinee
SHIFT SUMMARY:
LOG:

 

H1 General
nutsinee.kijbunchoo@LIGO.ORG - posted 16:06, Sunday 30 April 2017 (35922)
Ops Eve Shift Transition

Wind's reaching 40 mph but the IFO is still locked.

H1 General
edmond.merilh@LIGO.ORG - posted 15:50, Sunday 30 April 2017 (35920)
"TypeError while runnur test: IFO_STATUS"

Got this message, in RED, twice on Verbal Alarms computer. 21:02 & 22:35UTC

H1 CDS (DetChar)
edmond.merilh@LIGO.ORG - posted 13:28, Sunday 30 April 2017 - last comment - 14:13, Sunday 30 April 2017(35916)
H1IOPSUSAUX2 Crash

@ 20:19 the H1SUSAUX FE crashed

Comments related to this report
edmond.merilh@LIGO.ORG - 13:46, Sunday 30 April 2017 (35917)

13:40UTC Spoke to Dave. He had me reset the FE computer. Everything came back up rather quickly and I was able to get back to Observing by 13:45UTC

edmond.merilh@LIGO.ORG - 13:56, Sunday 30 April 2017 (35918)
Images attached to this comment
david.barker@LIGO.ORG - 14:13, Sunday 30 April 2017 (35919)

This was a good test for Tuesday's planned reboots. We don't need to power cycle the IO Chassis when rebooting the computer. Since the computer's general core is locked up, the only way to reboot it is to press the front panel Reset button (a power cycle of the computer is not needed).

Looking at the up-times, we only need to reboot those with times close to or beyond 208 days, which means we don't have to immediately reboot:

h1psl0          up 133 days,
h1susauxh2      just rebooted
h1susauxh34     up 120 days
h1oaf0          up 166 days
h1susex         booted Friday
h1seiex         booted Thursday
h1iscex         up 179 days

We have the option of deferring the h1psl0 reboot and the others until the 5/8 vent.

H1 General
edmond.merilh@LIGO.ORG - posted 08:13, Sunday 30 April 2017 (35914)
Shift Transition - Day

TITLE: 04/30 Day Shift: 15:00-23:00 UTC (08:00-16:00 PST), all times posted in UTC
STATE of H1: Observing at 65Mpc
OUTGOING OPERATOR: Corey
CURRENT ENVIRONMENT:
    Wind: Calm
    Primary useism: 0.09 μm/s
    Secondary useism: 0.25 μm/s
QUICK SUMMARY: ops lazy script isn't doing transition (-t). It didn't work for me yesterday either. The End (-e) funcion worked for me yesterday.

I ran a2l after seeing the measurement that Corey took just before he left

 
 
 
LHO General
corey.gray@LIGO.ORG - posted 08:10, Sunday 30 April 2017 (35913)
Owl Operator Summary

TITLE: 04/30 Owl Shift: 07:00-15:00 UTC (00:00-08:00 PST), all times posted in UTC

STATE of H1: Observing at 64Mpc

INCOMING OPERATOR: Ed

SHIFT SUMMARY:

Nice quiet shift.  H1 locked for over 38.5hrs.  L1 was under tornado warning.

H1 SEI (SEI)
corey.gray@LIGO.ORG - posted 07:41, Sunday 30 April 2017 (35912)
Earthquake Report: 5.2 New Zealand

5.2 Magnitude EQ from New Zealand originating at 12:22utc.

LHO General
corey.gray@LIGO.ORG - posted 04:58, Sunday 30 April 2017 (35911)
Mid-ish Shift Summary

Nice steady range (with a few glitches). Currently in a wind lull. H1 locked for over 35.5 hrs.

(I think it's lunch time.)

LHO General
corey.gray@LIGO.ORG - posted 00:51, Sunday 30 April 2017 (35910)
Transition To OWL

TITLE: 04/30 Owl Shift: 07:00-15:00 UTC (00:00-08:00 PST), all times posted in UTC
STATE of H1: Observing at 65Mpc
OUTGOING OPERATOR: Nutsinee
CURRENT ENVIRONMENT:
    Wind: 5mph
    Primary useism: 0.01 μm/s
    Secondary useism: 0.10 μm/s (at 50 percentile)
QUICK SUMMARY:

H1 locked for 3.5hrs & useism took a tiny step up 8hrs ago.

H1 General
nutsinee.kijbunchoo@LIGO.ORG - posted 00:41, Sunday 30 April 2017 (35909)
Ops Eve Shift Summary

TITLE: 04/30 Eve Shift: 23:00-07:00 UTC (16:00-00:00 PST), all times posted in UTC

STATE of H1: Observing at 62Mpc

INCOMING OPERATOR: Corey

SHIFT SUMMARY: Quiet shift. Been locked and observing for 31.5 hours. No issue to report.

H1 General
nutsinee.kijbunchoo@LIGO.ORG - posted 20:14, Saturday 29 April 2017 (35908)
Ops Eve Midshift Summary

Been locked and observing for almost 27 hours. No issue to report. LLO is down due to >20mph wind (at least that's what I gather from their alog).

H1 General
edmond.merilh@LIGO.ORG - posted 16:01, Saturday 29 April 2017 (35903)
Shift Summary - Day

TITLE: 04/29 Day Shift: 15:00-23:00 UTC (08:00-16:00 PST), all times posted in UTC
STATE of H1: Aligning
INCOMING OPERATOR: Nutsinee
SHIFT SUMMARY:

Basically quiet shift. There were a few tours through the control room. I ran the a2l script at the beginning ofthe shift while Livingston was down and the end of the shift when Livingston dropped out.
LOG:

15:30 big glitch

16:40 big glitch

19:16 I just noticed two guys with backpack sprayers, tending the grounds on the walkway between the Staging Building and the LSB.

H1 General
edmond.merilh@LIGO.ORG - posted 15:56, Saturday 29 April 2017 - last comment - 16:04, Saturday 29 April 2017(35906)
H1 a2l alignment while Livingston is down

22:55UTC YAW out by approx .75

Comments related to this report
edmond.merilh@LIGO.ORG - 16:04, Saturday 29 April 2017 (35907)

23:03 Intention Bit Undisturbed

H1 CDS
david.barker@LIGO.ORG - posted 08:58, Saturday 29 April 2017 - last comment - 09:19, Sunday 30 April 2017(35901)
update on possible cause of EX computer lock-ups

There is a write up of the bug which is possibly causing our front end issues here

I confirmed that our 2.6.34 kernel source on h1boot does have the arch/x86/include/asm/timer.h file with the bug.

To summarize, there is a low level counter in the kernel which has a wrap around bug with kernel versions 2.6.32 - 2.6.38. The bug causes it to wrap around after 208.499 days, and many other kernel subsystems assume this could never happen. 

At 06:10 Thursday morning h1seiex partially locked-up. At 05:43 Friday morning h1susex partially locked-up. As of late Wednesday night, all H1 front ends have been running for 208.5 days. The error messages on h1seiex and h1susex consoles show timers which had been reset Wednesday night, within two hours of each other.

We are going on the assumption that the timer wrap-around has put the front end computers in a fragile state where lock-ups may happen. We don't know why only two computers at EX have seen the issue, or why around 6am, or why one day apart. Nothing happened around 6am this morning.

I am filing a work permit to reboot all H1 front end computers and DAQ computers which are running kernels with this bug.

Comments related to this report
david.barker@LIGO.ORG - 09:15, Saturday 29 April 2017 (35902)DAQ

attached is a list of 2.6.3x kernel machine and their runtime. Looks like h1nds0 will wrap-around Sunday afternoon.

 

Images attached to this comment
david.barker@LIGO.ORG - 09:19, Sunday 30 April 2017 (35915)

h1nds0's timer will wrap around to zero about 11pm Sunday local time.

Displaying reports 51221-51240 of 86130.Go to page Start 2558 2559 2560 2561 2562 2563 2564 2565 2566 End