40m QIL Cryo_Lab CTN SUS_Lab TCS_Lab OMC_Lab CRIME_Lab FEA ENG_Labs OptContFac Mariner WBEEShop
  40m Log, Page 327 of 341  Not logged in ELOG logo
ID Date Author Type Category Subjectup
  7780   Tue Dec 4 10:44:36 2012 SteveUpdatePEMshort power outage

 

CALIFORNIA INSTITUTE OF TECHNOLOGY

                 FACILITIES MANAGEMENT

 

**PLEASE POST**

 

 

Building:         Campus

 

Date:             Thursday October 04,2012

 

This morning at 2:17 a.m. much of the City of Pasadena including our Campus experienced a electric power sag of short duration, approximately 1/10 of a second. The cause was a fault on one of Pasadena’s 17KV circuits. Some sensitive equipment have been impacted.

                 

Contact:          Mike Anchondo x-4999

  14914   Mon Sep 30 13:20:55 2019 aaronUpdateIOOshot noise measurement

I wanted to measure the RF transimpedance of the WFS heads, as outlined above.

Summary: Measurement is not done.

Details:

  • closed the PSL shutter
  • taped over the WFS 2 opening with frosted scotch tape
  • illuminated the QPD with an incandescent flashlight.
    • All of the D batteries were close to dead, so it seemed dimmer than usual
  • Observed the WFS2 segment 1 RF spectrum on the Agilent, but saw no difference between the spectrum with and without the flashlight. Must need a brighter light, and possibly also better alignment.
  • Needed to skype someone and pass off the IFO to gautam, so I untaped the QPD, returned the appropriate LEMO connector, and opened the PSL shutter.
  5328   Wed Aug 31 11:37:56 2011 steveUpdateGreen Lockingshutter in place at east end

Uniblitz mechanical shutter installed in the green beam path at ETMY-ISCT  The remote control cable has not been connected.

  9319   Wed Oct 30 14:58:44 2013 SteveHowToGeneralshutting down a computer

 

 We have tried to ssh into c1iscey yesterday morning. It just did not work. We have just tried it again (now) and it did work.

Lesson learned: always shut down the computer from a TERMINAL Do not turn it off by the manual power switch.

  5598   Mon Oct 3 04:43:03 2011 kiwamuUpdateLSCsideband-resonance PRMI locked

My goal of today was to lock PRMI without using AS55 and it is 50% successful.

The sideband-resonance PRMI (SB-PRMI) was locked with REFL33_I and REFL55_Q for the PRCL and MICH control respectively.

The carrier-resonance PRMI wasn't able to be locked without AS55.

     (it looked no clean MICH signals at the REFL ports.)

 


(Motivation)

 The motivation of not to use AS55 came from the suspicion that AS55 was injecting some noise into MICH (#5595).

So I wanted to try another RFPD to see if it helps the stability or not. 

 

(locking activity)

  The lock of SB-PRMI was quite stable so that it stayed locked more than 30 minutes (it ended because I turned off the servos.)

Then I briefly tried DRMI while PRCL and MICH kept locked by the same control loops, namely REFL33_I and REFL55_Q.

The lock of MICH and PRCL looked reasonably robust against the SRCL fringes, but wasn't able to find a good signal for SRCL.

I think I am going to try locking DRMI tomorrow.

 - - settings

    Demod phase for REFL55 = -45.3 deg

    Demod phase for REFL33 = -14.5 deg

    Whitening gain for REFL55 = 4 (12 dB)

    Whitening gain for REFL33 = 10 (30 dB)

    MICH gain = 100

    PRCL gain = 8

 

(misc.)

 + I removed an iris on the ITMY table because it was in the way of POY. See the picture below.

ITMYtable.png

 

  + I found that burtrestore for the ETMX DC coil forces were not functional.

  => currently ETMX's "restore" and "mislalign" buttons on the C1IFO_ALIGN screen are not working.

  => According to the error messages, something seemed wrong on c1auxex, which is a slow machine controlling the DC force.

  4361   Sat Feb 26 02:33:16 2011 kiwamuUpdateGreen Lockingsidebands on beatnote

The power ratio of the beatnote signal vs. the 216kHz sideband has been measured.

The measured ratio was -55 dB, which is smaller by about 20 dB than Aidan's estimation.

To confirm this fact we should check the modulation depth of the end PDH somehow.

 

The below is a picture showing the sidebands around the beatnote locked at 66.45 MHz.

Other than the +/-216 kHz sidebands, we can see some funny peaks at +/- 50 kHz and +/-150 kHz

I wonder if they come from the servo oscillation of the MC servo whose UGF is at 24 kHz.  We can check it by unlocking the MC.

beat_note.png

Quote: #4351 by Aidan

So, on the vertex PD, the power of the 80MHz +/-200kHz sidebands should be around sqrt(0.15)*0.05 = 0.02 = 2% of the 80MHz beatnote.

Once we get the green and IR locked to the arm again, we're going to look for the sidebands around the beatnote.

  4362   Sun Feb 27 09:43:59 2011 AidanUpdateGreen Lockingsidebands on beatnote

Can we set up a fiber-PD on the end table to look at the beat between the "end laser IR beam" and the "PSL IR beam fiber-transmitted end beam"? 

We should see the same thing on that PD that we see on the green PD (plus any fiber noise and I'm not really sure how much that'll be off the top of my head). If we unlock the lasers from the arm cavity then the free-running noise of the lasers wrt to each other will probably swamp the 50kHz and 150kHz signals. Maybe we could lock the end laser to the free-running PSL by demodulating the beat note signal from the fiber-PD and then we could look for the extra sidebands in the IN-LOOP signal. Then we could progressively lock the PSL to the MC and arm cavity and see if the sidebands appear on the fiber-PD at some point in that process. 

It's possible that the 216kHz drive of the PZT on the Innolight is somehow driving up some sub-harmonics in the crystal. I think this is unlikely though: if you look at Mott's measurements of the Innolight PZT response, there are no significant PM resonances at 50 or 150kHz.

 

Quote:

Other than the +/-216 kHz sidebands, we can see some funny peaks at +/- 50 kHz and +/-150 kHz.

 

Quote: #4351 by Aidan

So, on the vertex PD, the power of the 80MHz +/-200kHz sidebands should be around sqrt(0.15)*0.05 = 0.02 = 2% of the 80MHz beatnote.

Once we get the green and IR locked to the arm again, we're going to look for the sidebands around the beatnote.

 

  4363   Sun Feb 27 13:09:56 2011 ranaUpdateGreen Lockingsidebands on beatnote

When Koji and I were massaging the MC, we noticed that the oscillations were at 48.5 kHz. They were pretty huge and are probably what you're seeing on the beat. My guess is that they are the PZT resonances of the PSL 2W NPRO; we need to put a notch in the FSS box - it still has the notch from the old NPRO.

  2757   Thu Apr 1 20:29:02 2010 HartmutUpdateGreen Lockingsimple PD test circuit

I made a simple PD test circuit which may allow to test PD response up to few 100MHz.

Its not for low noise, only for characterising PD response.

Here is the circuit:

The 2 capacitor values (for bypassing) are kind of arbitrary, just what I found around

(one medium, one small capacity). Could be improved by better RF types (e.g. Mica).

The PD type has no meaning. I put in the Centronic 15-T5 for a start.

The bias can be up to 20V for this diode.

The signal appears across R1. It is small, to make a large bandwidth.

R2 is just for slightly decoupling the signal from the following RF amplifier.

The wire into the RF amplifier is short (~cm). And the amplifier is supposed to have 50 Ohm

input impedance.

I use a mini circuits ZFL 500 here.

power supply for this is 15V.

pdtest.png

  3048   Thu Jun 3 22:33:31 2010 valeraSummaryCDSsimulated plant work

 I put matlab files and a summary into the 40m wiki for the fitting of the 40m Optickle transfer functions and generating digital filters for the simulated plant:

http://lhocds.ligo-wa.caltech.edu:8000/40m/Generating_DOF-%3EPD_digital_filters_based_on_Optickle_modeling

The filters are not loaded yet. Joe and Alex will make a rcg code to make a matrix of filters (currently 5x15=75 elements) which will enable the simulated plant tf's.

Joe and I tried to put a signal through the DARM loop but the signal was not going through the memory location in the scx part of the simulated plant.

 

Edit by Joe:

I was able to track it down to the spx model not running properly.  It needed the Burt Restore flag set to 1.  I hadn't done that since the last rebuild, so it wasn't actually calculating anything until I flipped that flag.  The data is now circulating all the way around.  If I turn on the final input (the same one with the initial 1.0 offset), the data circulates completely around and starts integrating up.  So the loop has been closed, just without all the correct filters in.

  6874   Tue Jun 26 01:30:13 2012 yutaSummaryGreen Lockingsimultaneous hold and release of the arm (aka two arm ALS)

To get the feeling of the master of IFO, I;

1. Stabilized both arm length using ALS.

2. Ran findIRresonance.py for both arms and find what offset gives me IR resonances.

3. Holded X arm to IR resonance, holded Y arm to IR resonance, and released both arms.

Below is the time series data of what I did.
ALSboth20120625_2withAS.png


Issues:
 - Currently ALS is not stable enough. It only stays for about few minutes. I think it is because of the bad alignment of green from each end.
 - We can't tell end green frequency is higher or PSL green frequency is higher. So, the sign of the servo filter sometimes flips.
 - Wobbliness of X end green transmission beam spot was from the ETMX oplev. When the oplev servo is on, it got more wobbly when X end table is opened. But when the oplev servo was off, wobbliness was same even if the presence of air flow.
 - MICH contrast in plot above seems like it somehow got better when two arms are at resonance by ALS. I think this is not real because reflection from both arms at AS port was not well aligned and beam was clipped. Koji and I measured contrast of FPMI and MI(ETMs misalined), and they were 99.6 % and 99.9 % respectively. Beam clipping seems like it comes from some where between BS to AS port. We need to figure out where and fix this.

Things need to be done to make ALS more concrete:
 - Align Y end green beam!
 - Look into Y end green frequency servo
 - How do we hand-off servo using ALS to IR lock?
 - Noise budgeting for new phase tracker scheme
 - Linearity check of the beat box and phase tracker

  4391   Wed Mar 9 17:29:11 2011 steveSummaryVACsingle O-ring protection

We have one single O-ring on the 40m vacuum envelope. It is on the OOC west side, facing the AP table. This O-ring has to be protected from the force of this

door. There should be 3 shims  ~120 degrees apart to carry the full load, so it is not the O-ring that is getting squashed.

This morning I found only one of these shims in place.

Attachment 1: so1.jpg
so1.jpg
Attachment 2: P1070458.JPG
P1070458.JPG
  6046   Tue Nov 29 22:54:49 2011 DenUpdatedigital noisesingle precision

 In the previous elog we've proved that signals C1:SUS-MC3_LSC_EXCMON and  C1:SUS-MC3_LSC_OUTMON are in single precision. Now we try to understand if the precision is lost when the signals enter dtt tools or in the online code. For this measurement we just switch on one of the filters between the signals. By this we add mathematical operations inside the online code. If double precision is used there, then we'll see the same error as before, because the real error is still very small (~10-15) and dtt tools add this 10-7 error. But if the digital error will increase then no matter what kind of precision use dtt, online code uses single precision. At least at some points.

Test 1.   cheby1("LowPass",6,1,12) 

filter_coherence.jpg

 

Test 2.  cheby1("Lowpass",2,0.1,3)

chby1_coherence.jpg

 Test 3. butter("LowPass",8,10)

butter8_coherence.jpg

Test 4.  butter("LowPass",2,10)

butter2_coherence.jpg

We can see that coherence become worse. And longer filter - more digital error. This means that single precision is used in calculations.

 

  6717   Wed May 30 18:16:44 2012 JamieUpdateLSCskeleton of new c1cal calibration model created

[Jamie, Xavi Siemens, Chris Pankow]

We built the skeleton of a new calibration model for the LSC degrees of freedom.  I named it "c1cal".  It will run on the c1lsc FE machine, in CPU slot 4, and has been given DCUID 50.

Right now there's not much in the model, just inputs for DARM_ERR and DARM_CTRL, filters for each input, and the sum of the two channels that is h(t).

Tomorrow we'll extract all the needed signals from c1lsc, and see if we can generate something resembling a calibrated signal for one of the IFO DOFs.

 

  617   Tue Jul 1 21:27:27 2008 robHowToComputer Scripts / Programsslider twiddling after reboot

Sometimes after we reboot the front-end machines, some of the hardware gets stuck in an unknown state. We generally fix this by twiddling EPICS settings, which refresh the hardware somehow and put it into a known state. I've started a script (slider_twiddle) which we can just run after reboots to do this for us. Right now it just has the QPD whitening gain settings. As we find more stuff, we can add to it. It's in $SCRIPTS/Admin/.
  9824   Thu Apr 17 16:59:45 2014 jamieUpdateCDSslightly more successful attempt to get Dolphin working on c1ioo

So it turns out that the card that Rolf had given me was not a Dolphin host adapter after all.  He did have an actual host adapter board on hand, though, and kindly let us take it.  And this one works!

I installed the new board in c1ioo, and it recognized it.  Upon boot, the dolphin configuration scripts managed to automatically recognize the card, load the necessary kernel modules, and configure it.  I'll describe below how I got everything working.

However, at some point mx_stream stopped working on c1ioo.  I have no idea why, and it shouldn't be related to any of this dolphin stuff at all.  But given that mx_stream stopped working at the same time the dolphin stuff started working, I didn't take any chances and completely backed out all the dolphin stuff on c1ioo, including removing the dolphin host adapter from the chassis all together.  Unfortunately that didn't fix any of the mx_stream issues, so mx_stream continues to not work on c1ioo.  I'll follow up in a separate post about that.  In the meantime, here's what I did to get dolphin working on c1ioo:

c1ioo Dolphin configuration

To get the new host recognized on the Dolphin network, I had to make a couple of changes to the dolphin manager setup on fb.  I referenced the following page:

https://cdswiki.ligo-la.caltech.edu/foswiki/bin/view/CDS/DolphinHowTo

Below are the two patches I made to the dolphin ("dis") config files on fb:

--- /etc/dis/dishosts.conf.bak    2014-04-17 09:31:08.000000000 -0700
+++ /etc/dis/dishosts.conf    2014-04-17 09:28:27.000000000 -0700
@@ -26,6 +26,8 @@
 ADAPTER:  c1sus_a0 8 0 4
 HOSTNAME: c1lsc
 ADAPTER:  c1lsc_a0 12 0 4
+HOSTNAME: c1ioo
+ADAPTER:  c1ioo_a0 16 0 4
 
 # Here we define a socket adapter in single mode.
 #SOCKETADAPTER: sockad_0 SINGLE 0

--- /etc/dis/networkmanager.conf.bak    2014-04-17 09:30:40.000000000 -0700
+++ /etc/dis/networkmanager.conf    2014-04-17 09:30:48.000000000 -0700
@@ -39,7 +39,7 @@
 # Number of nodes in X Dimension. If you are using a single ring, please
 # specify number of nodes in ring.
 
--dimensionX 2;
+-dimensionX 3;
 
 # Number of nodes in Y Dimension.

I then had to restart the DIS network manager to see these changes take affect:

$ sudo /etc/init.d/dis_networkmgr restart

I then rebooted c1ioo one more time, after which c1ioo showed up in the dxadmin GUI.

At this point I tried adding a dolphin IPC connection between c1als and c1lsc to see if it worked.  Unfortunately everything crashed every time I tried to run the models (including models on other machines!).  The problem was that I had forgotten to tell the c1ioo IOP (c1x03) to use PCIe RFM (i.e. Dolphin).  This is done by adding the following flag to the cdsParamters block in the IOP:

pciRfm=1

Once this was added, and the IOP was rebuilt/installed/restarted and came back up fine.  The c1als model with the dolphin output also came up fine.

However, at this point I ran into the c1ioo mx_stream problem and started backing everything out.

 

  3728   Fri Oct 15 15:32:35 2010 josephbUpdateCDSslow DAQ channels added

I added all the _OUT16, _INMON, _EXCMON channels associated with the BS, ITMX, and ITMY channels in SUS_SLOW.ini in the /opt/rtcds/caltech/c1/chans/daq directory.  Similarly, I added the channels associated with MC1, MC2, and MC3 to MCS_SLOW.ini and those associated with PRM and the SRM to RMS_SLOW.ini. 

Lines pointing to these files were then added to the /opt/rtcds/caltech/c1/target/fb/master file and the frame builder restarted.  It took about 4 tries before the frame builder stayed up using ./daqd -c ./daqdrc.

I generated the .ini files with a python script.  Its at /opt/rtcds/caltech/c1/scripts/daqscripts/create_inmon_out16_daq_ini.py. It checks the C0EDCU.ini to find what slow channels already exist, then goes through a medm directory given at the command line looking for all _OUT16, _INMON, and _EXCMON channels, adding them if they aren't already accounted for, and then writing out the new file to the location of your choice.

  3732   Mon Oct 18 08:44:36 2010 ranaUpdateCDSslow DAQ channels added

Why EXCMON? I think that we should modify the filter coefficients so that the decimation filter that makes OUT16 is a 2nd order Butterworth at 1 Hz instead of whatever bogus thing it is now. Then we can delete the EXCMON from the DAQ and add either OUTPUT or OUTMON. That way we will have a low frequency signal (OUT16) and a sort of aliased RMS (OUTMON).

  6915   Thu Jul 5 01:20:58 2012 yutaSummaryCDSslow computers, 0x4000 for all DAQ status

ALS looks OK. I tried to lock FPMI using ALS, but I feel like I need 6 hands to do it with current ALS stability. Now I have all computers being so slow.

It was fine for 7 hours after Jamie the Great fixed this, but fb went down couple times and DAQ status for all models now shows 0x4000. I tried restarting mx_stream and restarting fb, but they didn't help.

  12754   Wed Jan 25 14:30:20 2017 gautamUpdateCDSslow machine bootfest

[gautam, lydia]

We rebooted c1psl, c1iscaux and c1aux which were all showing the typical symptom of responding to ping but not to telnet (and also blanked out epics fields on the MEDM screens). Keyed all these crates.

Restored burt snapshots for c1psl, PMC locked fine, and IMC is also locked now.

Johannes forgot to elog this yesterday, but he rebooted c1susaux following the usual procedure to avoid getting ITMX stuck. 

 

  12762   Fri Jan 27 17:07:52 2017 LydiaUpdateCDSslow machine bootfest

Rebooted c1iscaux, c1auxex and c1auxey which were all not reponding to telnet. The watchdogs for the ETMs were turned off and then I keyed all 3 crates. All slow machines are reponding to telnet now. Both green lasers locked to the arms so I didn't do any burt restore.

  12803   Mon Feb 6 15:18:08 2017 gautamUpdateCDSslow machine bootfest

Had to reboot c1psl, c1susaux, c1auxex, c1auxey and c1iscaux today. PMC has been relocked. ITMX didn't get stuck. According to this thread, there have been two instances in the last 10 days in which c1psl and c1susaux have failed. Since we seem to be doing this often lately, I've made a little script that uses the netcat utility to check which slow machines respond to telnet, it is located at /opt/rtcds/caltech/c1/scripts/cds/testSlowMachines.bash.

The script can be executed by ./testSlowMachines.bash.

  13012   Thu May 25 12:22:59 2017 gautamUpdateCDSslow machine bootfest

After ~3months without any problems on the slow machine front, I had to reboot c1psl, c1susaux and c1iscaux today. The control room StripTool traces were not being displayed for all the PSL channels so I ran testSlowMachines.bash to check the status of the slow machines, which indicated that these three slow machines were dead. After rebooting the slow machines, I had to burt-restore the c1psl snapshot as usual to get the PMC to lock. Now, both PMC and IMC are locked. I also had to restart the StripTool traces (using scripts/general/startStrip.sh) to get the unresponsive traces back online.

Steve tells me that we probably have to do a reboot of the vacuum slow machines sometime soon too, as the MEDM screen for the Vacuum indicator channels are unresponsive.

Quote:

Had to reboot c1psl, c1susaux, c1auxex, c1auxey and c1iscaux today. PMC has been relocked. ITMX didn't get stuck. According to this thread, there have been two instances in the last 10 days in which c1psl and c1susaux have failed. Since we seem to be doing this often lately, I've made a little script that uses the netcat utility to check which slow machines respond to telnet, it is located at /opt/rtcds/caltech/c1/scripts/cds/testSlowMachines.bash.

 

 

  13028   Thu Jun 1 15:37:01 2017 gautamUpdateCDSslow machine bootfest

Steve alerted me that the IMC wouldn't lock. Reboots for c1susaux, c1iool0 today. I tried using the reset button instead of keying the crates. This worked for c1iool0, but not for c1susaux. So I had to key the latter crate. The machine took a good 5-10 minutes before coming back up, but eventually it did. Now IMC locks fine.

  13059   Mon Jun 12 10:34:10 2017 gautamUpdateCDSslow machine bootfest

Reboots for c1susaux, c1iscaux, c1auxex today. I took this opportunity to squish the Sat. Box. Cabling for MC2 (both on the Sat box end and also the vacuum feedthrough) as some work has been recently ongoing there, maybe something got accidently jiggled during the process and was causing MC2 alignment to jump around.

Relocked PMC to offload some of the DC offset, and re-aligned IMC after c1susaux reboot. PMC and IMC transmission back to nominal levels now. Let's see if MC2 is better behaved after this sat. box. voodoo.

Interestingly, since Feb 6, there were no slow machine reboots for almost 3 months, while there have been three reboots in the last three weeks. Not sure what (if anything) to make of that.

  13069   Fri Jun 16 13:53:11 2017 gautamUpdateCDSslow machine bootfest

Reboots for c1psl, c1iool0, c1iscaux today. MC autolocker log was complaining that the C1:IOO-MC_AUTOLOCK_BEAT EPICS channel did not exist, and running the usual slow machine check script revealed that these three machines required reboots. PMC was relocked, IMC Autolocker was restarted on Megatron and everything seems fine now.

 

  13096   Wed Jul 5 16:09:34 2017 gautamUpdateCDSslow machine bootfest

Reboots for c1susaux, c1iscaux today.

 

  13273   Wed Aug 30 10:54:26 2017 gautamUpdateCDSslow machine bootfest

MC autolocker and FSS loops were stuck because c1psl was unresponsive. I rebooted it and did a burtrestore to enable PSL locking. Then the IMC locked fine.

c1susaux and c1iscaux were also unresponsive so I keyed those crates as well, after taking the usual steps to avoid ITMX getting stuck - but it still got stuck when the Sat. Box. connectors were reconnected after the reboot, so I had to shake it loose with bias slider jiggling. This is annoying and also not very robust. I am afraid we are going to knock the ITMX magnets off at some point. Is this problem indicative of the fact that the ITMX magnets were somehow glued on in a skewed way? Or can we make the situation better by just tweaking the OSEM-holding fixtures on the cage?

In any case, I've started listing stuff down here for things we may want to do when we vent next.

 

  13297   Tue Sep 5 23:02:37 2017 gautamUpdateCDSslow machine bootfest

MC autolocker was not working - PCdrive was railed at its upper rail for ~2 hours judging by the wall StripTool trace. I tried restarting the init processes on megatron, but that didn't fix the problem. The reason seems to have been related to c1iool0 failing - after keying the crate, autolocker came back fine and MC caught lock almost immediately.

Additionally, c1susaux, c1auxex,c1auxey and c1iscaux are also down. I'm not planning on using the IFO tonight so I am not going to reboot these now.

 

  13355   Tue Oct 3 19:39:10 2017 gautamUpdateCDSslow machine bootfest

Eurocrate key turning reboots for c1susaux, c1auxex,c1auxey, c1iscaux, c1iscaux2 and c1aux. Usual precautions were taken for ITMX. Did burtrestore for c1iscaux andc1iscaux2  in order to restore the LSC PD whitening gains.


Un-related to this work: input pointing into PMC was tweaked as the PMC_REFL spot was pretty bright.

  13360   Thu Oct 5 11:46:15 2017 gautamUpdateCDSslow machine bootfest

MC Autolocker was umnhappy because c1iool0 was unresponsive and hence it couldn't write to the "C1:IOO-MC_AUTOLOCK_BEAT" channel. I keyed the crate and IMC locked almost immediately. I'm moving this channel into the RTCDS model as we did for the IFO_STATE EPICS channel so that the autolocker isn't dependant on c1iool0 (which was the whole point of migrating the IFO-STATE variable anyways). I also commented out all of these channels in /cvs/cds/caltech/target/c1iool0/autolocker.db so that there aren't duplicate channels.

Quote:

Eurocrate key turning reboots for c1susaux, c1auxex,c1auxey, c1iscaux, c1iscaux2 and c1aux. Usual precautions were taken for ITMX. Did burtrestore for c1iscaux andc1iscaux2  in order to restore the LSC PD whitening gains.


Un-related to this work: input pointing into PMC was tweaked as the PMC_REFL spot was pretty bright.

 

  13379   Thu Oct 12 14:42:45 2017 gautamUpdateCDSslow machine bootfest

Steve reported problems getting the X arm locked. Alignment sliders were inaccessible. Eurocrate key turning reboots for c1susaux, c1auxex,c1auxey, c1iscaux and c1aux. Usual precautions were taken for ITMX.

This is becoming a once-a-week thing sad.

  13399   Tue Oct 24 16:43:11 2017 SteveUpdateCDSslow machine bootfest

[ Gautam , Steve ]

c1susaux & c1iscaux were rebooted manually.

Quote:

Had to reboot c1psl, c1susaux, c1auxex, c1auxey and c1iscaux today. PMC has been relocked. ITMX didn't get stuck. According to this thread, there have been two instances in the last 10 days in which c1psl and c1susaux have failed. Since we seem to be doing this often lately, I've made a little script that uses the netcat utility to check which slow machines respond to telnet, it is located at /opt/rtcds/caltech/c1/scripts/cds/testSlowMachines.bash.

The script can be executed by ./testSlowMachines.bash.

 

  13518   Tue Jan 9 11:52:29 2018 gautamUpdateCDSslow machine bootfest

Eurocrate key turning reboots today morning for and c1susaux, c1auxey and c1iscaux. These were responding to ping but not telnet-able. Usual precautions were taken to minimize risk of ITMX getting stuck.

 

  13522   Wed Jan 10 12:24:52 2018 gautamUpdateCDSslow machine bootfest

MC autolocker got stuck (judging by wall StripTool traces, it has been this way for ~7 hours) because c1psl was unresponsive so I power cycled it. Now MC is locked.

  13558   Fri Jan 19 11:13:21 2018 gautamUpdateCDSslow machine bootfest

c1psl, c1susaux, and c1auxey today

Quote:

MC autolocker got stuck (judging by wall StripTool traces, it has been this way for ~7 hours) because c1psl was unresponsive so I power cycled it. Now MC is locked.

 

  13727   Wed Apr 4 16:23:39 2018 gautamUpdateCDSslow machine bootfest

[johannes, gautam]

It's been a while - but today, all slow machines (with the exception of c1auxex) were un-telnetable. c1psl, c1iool0, c1susaux, c1iscaux1, c1iscaux2, c1aux and c1auxey were rebooted. Usual satellite box unplugging was done to avoid ITMX getting stuck.

  13758   Wed Apr 18 10:44:45 2018 gautamUpdateCDSslow machine bootfest

All slow machines (except c1auxex) were dead today, so I had to key them all. While I was at it, I also decided to update MC autolocker screen. Kira pointed out that I needed to change the EPCIS input type (in the RTCDS model) to a "binary input", as opposed to an "analog input", which I did. Model recompilation and restart went smooth. I had to go into the epics record manually to change the two choices to "ENABLE" and "DISABLE" as opposed to the default "ON" and "OFF". Anyways, long story short, MC autolocker controls are a bit more intuitive now I think.

Attachment 1: MCautolocker_MEDM_revamp.png
MCautolocker_MEDM_revamp.png
  13812   Thu May 3 12:19:13 2018 gautamUpdateCDSslow machine bootfest

Reboot for c1susaux and c1iscaux today. ITMX precautions were followed. Reboots went smoothly.

IMC is shuttered while Jon does PLL characterization...


Now relocked.

  13925   Thu Jun 7 12:20:53 2018 gautamUpdateCDSslow machine bootfest

FSS slow wasn't running so PSL PZT voltage was swinging around a lot. Reason was that was c1psl unresponsive. I keyed the crate, now it's okay. Now ITMX is stuck - Johannes just told be about an un-elogged c1susaux reboot. Seems that ITMX got stuck at ~4:30pm yesterday PT. After some shaking, the optic was loosened. Please follow the procedure in future and if you do a reboot, please elog it and verify that the optic didn't get stuck.

Attachment 1: ITMX_stuck.png
ITMX_stuck.png
  13410   Mon Nov 6 11:15:43 2017 gautamUpdateCDSslow machine bootfest + IFO re-alignment

Eurocrate key turning reboots today morning for and c1susaux, c1auxex and c1auxey. Usual precautions were taken to minimize risk of ITMX getting stuck.

The IFO hasn't been aligned in ~1week, so I recovered arm and PRM alignment by locking individual arms and also PRMI on carrier. I will try recovering DRMI locking in the evening.

As far as MC1 glitching is concerned, there hasn't been any major one (i.e. one in which MC1 is kicked by such a large amount that the autolocker can't lock the IMC) for the past 2 months - but the MC WFS offsets are an indication of when smaller glitches have taken place, and there were large DC offsets on the MC WFS servo outputs, which I offloaded to the DC MC suspension sliders using the MC WFS relief script.

I'd like for the save-restore routine that runs when the slow machines reboot to set the watchdog state default to OFF (currently, after a key-turning reboot, the watchdogs are enabled by default), but I'm not really sure how this whole system works. The relevant files seem to be in the directory /cvs/cds/caltech/target/c1susaux. There is a script in there called startup.cmd, which seems to be the initialization script that runs when the slow machine is rebooted. But looking at this file, it is not clear to me where the default values are loaded from? There are a few "saverestore" files in this directory as well:

  • saverestore.sav
  • saverestore.savB
  • saverestore.sav.bu
  • saverestore.req

Are the "default" channel values loaded from one of these?

  13408   Mon Oct 30 11:15:02 2017 gautamUpdateCDSslow machine bootfest + vacuum snafu

Eurocrate key turning reboots today morning for c1psl and c1aux.c1auxex and c1auxey are also down but I didn't bother keying them for now. PSL FSS slow loop is now active again (its inactivity was what prompted me to check status of the slow machines).

Note that the EPCIS channels for PSL shutter are hosted on c1aux.But looks like the slow machine became unresponsive at some point during the weekend, so plotting the trend data for the PSL shutter channel would have you believe that the PSL shutter was open all the time. But the MC_REFL DC channel tells a different story - it suggests that the PSL shutter was closed at ~4AM on Sunday, presumably by the vacuum interlock system. I wonder:

  1. How does the vacuum interlock close the PSL shutter? Is there a non-EPICS channel path? Because if the slow machine happens to be unresponsive when the interlock wants to close the PSL shutter via EPICS commands, it will be unable to. The fact that the PSL shutter did close suggests that there is indeed another path.
  2. We should add some feature to the vacuum interlock (if it doesn't already exist) such that the PSL shutter isn't accidentally re-opened until any vacuum related issues are resolved. Steve was immediately able to identify that the problem was vacuum related, but I think I would have just re-opened the PSL shutter thinking that the issue was slow computer related.
  3138   Tue Jun 29 17:10:49 2010 steve, ranaUpdateVACslow pumpdown started

The folding crane was fixed and tested this morning by the NNN rigging company. Pictures will be posted by Steve in the morning.

Afterwards, the ITM-east door was installed, jam-nuts checked. No high voltage was on for the in-vac PZTs.

The annulus spaces were roughed down to 350mTorr by Roughing Pump RP1. For this operation, we removed the low flow valve from the RP1 line.

After the spaces came down to ~400 mTorr, we closed their individual valves.

Warning: The VABSSC1 and VABSSC0 valves are incorrect and misleadingly drawn on the Vacuum overview screen.

We then:

  1. Closed V6 (valve between RP1 and the annulus line).
  2. Turned off RP1 from the MEDM screen.
  3. Installed the soft -starting butterfly valve.
  4. Turned on RP1.
  5. Opened V3.
  6. Closed VV1 (at the last minute - this is a vent valve and must be checked before each pumpdown)
  7. and pumpdown was started with a 3/4 turn opening of manual valve RV1.

Our idea is to have a much slower pumpdown this time than the last time when we had a hurricane kick up the dust. Looks like it worked, but next time we should do only 1/2 turn.

  3140   Tue Jun 29 23:49:18 2010 steve, ranaUpdateVACslow pumpdown started

Untitled.png

The pumpdown started at 4 PM (2300 UTC). At 10 PM, we (Jenne, Jan, and I) opened up the RV1 valve to full open. That's the second inflection point in the plot.

  3141   Wed Jun 30 00:45:26 2010 ranaUpdateVACslow pumpdown stopped for the night

As per Steve's instructions, at 12:43 AM, I used the following steps to stop the pumpdown until the morning:

  1. Close RV1 using the 'steering-wheel' wrench.
  2. Close V3.
  3. Turn OFF RP1.
  4. Disconnect RP1 hose at the plastic disconnect attached to the slow-start throttle valve.
  4089   Wed Dec 22 17:24:50 2010 steveConfigurationVACslow pumpdown at 12 hours

We have reached 200 Torr at 12 hours of slow pumping speed. Kiwamu stopped the pumping for 11 hrs last night .and I restarted it this morning.

Now RV1 is fully open with butterfly valve in place  and the second roughing  pump RP3 was just turned on.

 

How to stop pumping:

1,  close RV1 manual valve with torque wheel

2, close V3

3, turn off roughing pumps RP1 & RP3

4, disconnect metal hose connection to butterfly valve

Attachment 1: slowpd12h.jpg
slowpd12h.jpg
  4094   Thu Dec 23 13:30:09 2010 steveConfigurationVACslow pumpdown complete

The pump down continued this morning by the removal of the butterfly valve. Two roughing pumps were used to reach 500 mTorr

The Maglev monitoring MEDM screen "Rotating" indicator is not working. It is on all times. Please look at Maglev controller monitor for real information.

Pump down is completed.           

Configuration: vacuum normal after 86 days at atm               CC1 = 1e-5 Torr                  

                      IFO is hungry for light (and maybe some goulash with a little paprikash too)

Attachment 1: slow2dpd.jpg
slow2dpd.jpg
  3148   Wed Jun 30 15:24:04 2010 steve,kiwamuUpdateVACslow pumpdown copmlete

Quote:

Untitled.png

The pumpdown started at 4 PM (2300 UTC). At 10 PM, we (Jenne, Jan, and I) opened up the RV1 valve to full open. That's the second inflection point in the plot.

 Atm 2 is showing the butterfly valve that closes down down the orifice at higher pressure to slow down the pumping speed.

 See elog entry #2573

 

Attachment 1: slowpd.jpg
slowpd.jpg
Attachment 2: butterfly.JPG
butterfly.JPG
  3331   Fri Jul 30 09:55:34 2010 steveConfigurationVACslow pumpdown has started

Bob and Steve closed BS chamber with the help of the manual Genie lift and the pump down started. The PSL shutter was closed and manual block was placed in the beam path. High voltage power supplies were checked to be off.

Pumping speed ~ 1 Torr/min was achieved at  1/8 of a turn opened roughing valve RV1

Attachment 1: slowpd5hr.jpg
slowpd5hr.jpg
  2573   Fri Feb 5 11:01:49 2010 steveConfigurationVACslow pumpdown valve

I have installed a slow start throttle valve in front of V3  This spring loaded valve will cut down on the flow at high pressures. There will be no more sand storme

and static built-up during pump down.

Attachment 1: slow.JPG
slow.JPG
ELOG V3.1.3-