40m QIL Cryo_Lab CTN SUS_Lab TCS_Lab OMC_Lab CRIME_Lab FEA ENG_Labs OptContFac Mariner WBEEShop
  40m Log, Page 313 of 339  Not logged in ELOG logo
ID Date Author Type Category Subjectup
  3987   Fri Nov 26 16:37:29 2010 kiwamuUpdatePhotospictures on PIcasa

 I uploaded some pictures taken in the last and this week. They are on the Picasa web albums.

 in vac work [Nov. 18 2010]

 in vac work [Nov 23 2010]

 CDS work [Nov 24 2010]

 DSC_2657_ss.jpg

  2118   Mon Oct 19 14:48:15 2009 rana, robSummaryElectronicspiezo jena measuring box
Attached is the schematic of the Piezo Jena driver measuring box made in a Pomona box:
                2.2 uF
In ----o-------- | | --------o-------- Out
       |                     |
       _                     |
       _  1uF                R  7.5 kOhms
       |                     |
       |                     |
      GND                   GND
The 1 uF cap is there to simulate the piezo and the 2.2 uF and 7.5k resistor ac couple the signal for the spectrum analyzer. They give a ~10 Hz corner frequency.
Attachment 1: PA160153.JPG
PA160153.JPG
Attachment 2: PA160151.JPG
PA160151.JPG
  13541   Fri Jan 12 18:08:55 2018 gautamUpdateGeneralpip installed on nodus

After much googling, I figured out how to install pip on SL7:

sudo easy_install pip

Next, I installed git:

sudo yum install git A

Turns out, actually, pip can be installed via yum using

sudo yum install python-pip
  3937   Wed Nov 17 02:53:41 2010 yutaUpdateIOOplaced new PRM to BS table

(Kiwamu, Yuta)

Background:
  Yesterday, we aligned the Faraday and the beam reached SM2 at BS table.
  Today, we placed a new PRM tower to BS table.

What we did:

  1. Moved IPPO, IPPOSSM1, IPPOSSM3, IPANGSM1, IPANGSM2 out from the BS chamber.

  2. Moved SRM tower(at PRM's place) to the ITMX chamber.

  3. Placed the new PRM tower at the BS chamber.

  4. Adjusted positions of the OSEMs for PRM and BS so that the sensor output can have roughly half of their maximum.

  5. Checked damping servo for PRM and BS. They were working and helped us when adjusting OSEM positions.

  6. Placed IPPO back and using SM2, made the beam hit PR2 at ITMX table.

  7. Aligned the PRM so that the reflected beam path overlaps the incident beam.
     We checked it by looking at MMT1.
     For the alignment, we used IFO align sliders(C1:SUS-PRM_PIT_COMM, YAW_COMM).
        To use them, we rebooted c1susaux.

Result:
  1. The new PRM tower is placed.

  2. OSEM sensor outputs for PRM and BS are;

(V) PRM BS
max current value max current value
ULSEN 1.72 1.006 1.50 0.757
URSEN 1.66 0.918 1.57 0.821
LRSEN 1.92 1.304 1.57 0.821
LLSEN 2.06 1.031 1.38 0.704
SDSEN 9.21 4.366 1.57 0.821

    We changed PRM aligning slider values, and they changed OSEM sensor outputs. We set the slider values to 0 when adjusting OSEM positions.

  4372   Thu Mar 3 00:12:52 2011 kiwamuUpdateGreen Lockingplan
Tomorrow's tasks
  - Auto noise budget (Jamie)
  - Demodulation phase adjustment (Kate)
  - Auto alignment for green (Joe/Kiwamu)
  - ADC connection for the X end green REFL_DC ( )
  - remote local boost for the X end green ( )
  - TDS stuff (Joe)
  - check harmonic distortions on the RF distribution box (Larisa/Koji)
  - connect the X end mechanical shutter to c1auxex (Steve)
  7371   Mon Sep 10 19:04:32 2012 ranaUpdateGeneralplan

 On Friday, Koji and I adjusted the beam pointing into the DRMI using the PZT yaw and found that the beam inside the DRMI (as seen on the AS camera) looked OK (not distorted too much).

So it seems that the issue seen before, namely that the DRMI resonant mode is very strange, is no longer true.

The camera image at the AS port still looks elliptical. So Jenne and Mike have started to make this beam round by adjusting the lenses.

Our plan now is:

1) Fix AS camera optics to get a round beam (single bounce off of ITMY).

2) Flash DRMI to make sure the beam at AS is still round.

3) Using the moveable Watec camera and Sensoray, get images of the spot on all DRMI mirrors with DRMI flashing. Use targets and rulers whenever possible to get quantitative measurements of the beam positions. (i.e. just saying "Oh, its pretty much in the center" is the Mickey Mouse approach to science)

4) Align all pickoff beams in this situation. Make sure there is no in vac clipping. Align IP POS and ANG using this input beam pointing.

5) Pump down.

 

  4421   Tue Mar 22 00:01:25 2011 kiwamuSummaryGreen Lockingplan for daytime tasks

Some tasks for the daytime tomorrow.

  * Beam profile measurements of the Y end laser  (Suresh / Bryan)

  * Taking care of CDS and the simulated plant (Jamie / Joe)

  * Reconnect the X end mechanical shutter to 1X9 (Kiwamu)

  * LPF for the X end temperature feedback (Larisa)

  4159   Fri Jan 14 20:37:00 2011 kiwamuHowToGreen Lockingplan for this month

 I summarized how we proceed our green locking in this month on the wiki.

Since step1 and 2 shown on the wiki are mostly done apparently, so we will move on to step 3-D and 3-E.

A short term target in the coming couple of days is to phase lock the VCO to the beat note.

green_plan.png

  4419   Mon Mar 21 16:49:11 2011 kiwamuSummaryGreen Lockingplan for this week

- Plan for this week

  * Intensity stabilization for the end green laser (Matt / Kiwamu)

  * Hand off the servo from Green to Red (Matt / Kiwamu)

  * Y end green locking (Suresh / Bryan) (rough schedule)

  * Reconnect the X end mechanical shutter to 1X9 (Kiwamu)

  * Connect the end DCPD signal to a DAC (done)

  * Make a LPF in a Pomona box for the temperature (Larisa)

  * Clean up and finalize the X end setup (Kiwamu)

  * Make a item lists for electronics. Order the electronics. (Aidan / Kiwamu)

  13825   Tue May 8 10:24:10 2018 KiraSummaryPEMplan for this week

Here are a few things I will be working on:

  • Design PCB boards for the heater circuit and temperature sensor circuits [by wednesday]
  • Order the front panel I've designed for the seismometer block [today]
  • [next week?] install the new Acromag when it comes
  4426   Wed Mar 23 00:51:47 2011 kiwamuUpdateGreen Lockingplan for tomorrow

  - Plan for tomorrow

    * Video cable session (I need ETMY_TRNAS) (team)

    * Characterization of the Y end laser  (Bryan / Suresh)

    * LPF for the X end laser temperature control (Larisa)

    * Frequency Divider  (Matt)

    * X end mechanical shutter (Kiwamu)

  13501   Wed Jan 3 18:00:46 2018 gautamUpdatePonderSqueezeplan of action

Notes of stuff we discussed @ today's meeting, and afterwards, towards measuring ponderomotive squeezing at the 40m.

  1. Displacement noise requirements
    • Kevin is going to see if we can measure any kind of squeezing on a short timescale by tuning various parameters.
    • Specifically, without requiring crazy ultra low current noise level for the coil driver noise.
  2. Investigate how much actuation range we need for lock acquisition and maintaining lock.
    • Specifically, for DARM.
    • We will measure this by having the arms controlled with ALS in the CARM/DARM basis.
    • Build up a noise budget for this, see how significant the laser noise contribution is.
  3. RC folding mirrors
    • In the present configuration, these are introducing ~2.5% RT loss in the RCs.
    • This affects PRG, and on the output side, measurable squeezing.
    • We want to see if we can relax the requirements on the RC folding mirrors such that we don't have to spend > 20 k$.
    • Specifically, consider spec'ing the folding mirror coatings to only have HR @1064 nm, and take what we get at 532 nm.
    • But still demand tolerances on RoC driven by mode-matching between the RCs and the arm cavities.
  4. ALS with Beat Mouth
    • Use the fiber coupled light from the ends to make the ALS signals.
    • Gautam will update diagram to show the signal chain from end-to-end (i.e. starting at AUX laser, ending at ADC input).
    • Make a noise budget for the same - preliminary analysis suggests a sensing noise floor of ~10 mHz/rtHz.

RXA:

  • For the ALS-DARM budget the idea is that we can do lock acquisition better, so we don't need to care about the acquisition reqs. i.e. we just need to set the ETM coil driver current range based on the DARM in-lock values.
    • To get the coil driver noise to be low enough to detect squeezing we need to use a ~10-15 kOhm series resistor.
    • We assume that all DAC and coil driver input noises can be sufficiently filtered.
    • We are assuming that we don't change the magnet sizes or the number of coil windings in the OSEMs.
    • The noise in the ITMs doesn't matter because we don't use them for any locking activity, so we can easily set the coil driver series resistors to 15 kOhm.
    • We will do the bias for the ETMs and ITMs using some HV circuit (not the existing ones on the coil driver boards) and doing the summation after the main coil driver series resistor. This HV bias module needs to handle the ~ (2 V / 400 Ohm) = 5 mA which is now used. This would require (5 mA) x (15 kOhm) = 60+ V drivers.
  • IF we can get away with doing the ALS beat note with just red (still using GREEN light from the end laser to lock to the arms from the ends), we will not have any requirements for the 532 nm transmission of any optics in the DRMI area.
    • Get some quotes for the new PR/SR mirrors having tight RoC tolerance, high R for 1064, and no spec for 532.
    • Check that the 1-way fiber noise for 1064 nm is < 100 mHz/rHz in the 50-1000 Hz band. If its more, explore putting better acoustic foam around the fiber run.
    • Improve the mode-matching of the IR beam into the fibers at the ends. We want >80% to reduce the noise do to scattering; we don't really care about the amount of light available in the PSL - this is just to reduce the IR-ALS noise.
  8746   Tue Jun 25 19:18:07 2013 gautamConfigurationendtable upgradeplan of action for PZT installation

  This entry is meant to be a sort of inventory check and a tentative plan-of-action for the installation of the PZT mounted mirrors and associated electronics on the Y-endtable. 

Hardware details:

  •  PZT mounts are cleaned and ready to be put on the end-tables.
  • The PZTs being used are PI S-330.20L Piezo Tip/Tilt Platforms. Each endtable requires two of these. The input channels have male single-lemo connectors. There are 3 channels on each tip/tilt platform, for tilt, yaw and a bias voltage.
  • The driver boards being used are D980323 Rev C. Each board is capable of driving 2 piezo tip/tilt platforms. I am not too sure of this but I think that the SMA female connector on these boards is meant to be connected with the bias voltage from our Kepco high-voltage power supplies. The outputs on these boards are fitted with SMB female connectors, while the piezo tip/tilt platforms have male single-lemo connectors. We will have to source cables with the appropriate connectors to run between the end-table and rack 1Y4 (see below). The input to these boards from the DAC will have to be made with a custom ribbon connector as per the pin out configuration given in the circuit drawing.
  • High-voltage power supply: KEPCO BHK 300-130 MG. This will supply the required 100V DC bias voltage to the piezo tip/tilts via the driver board. Since each board is capable of driving two piezos, we will only need one unit per end-table. The question is where to put these (photo attached). It doesn't look like it can be accommodated in 1Y4 (again photo attached) and the power cable the unit came with is only about 8ft long. If we put these under the end-tables, then we will need an additional long (~10m) cable to run from these to the driver boards at 1Y4 carrying 100 V. 
  •  We will need long (~10m by my rough measurement at the X and Y ends) cables to run from rack 1Y4 to the endtable to drive the piezos. These will have to be high-voltage tolerant (at least to 100V DC) and should have SMB male connectors at one end and female single-lemo connectors at the other. I have emailed 3 firms (CD International Technologies Inc., Stonewall Cables, and Fairview Microwave) detailing our requirements and asking for a quote and estimated time for delivery. We will need 6 of these, plus another cable with an SMA connector on one end and the other end open to connect the 100V DC bias voltage from the high voltage power supply to the driver boards (the power supply comes with a custom jack to which we can solder open leads). We will also possibly need ~3m long lemo-to-?(I need to check what the input connector for the data acquisition channels) cables for the monitoring channels, I am not sure if these are available, I will check with Steve tomorrow.

Other details:

  • I have attached a wiring diagram with the interconnects between various devices at various places and the type of connectors required etc. The error signal will the the transmitted green light from the cavity, and there is already a DQ channel logging this information, so nothing additional wiring is required to this end.
  • Jamie had detailed channel availability in elog 8580. I had a look at rack 1Y4, and there were free DAC channels available, but I am not sure as to which of the ones listed in the elog it corresponds to. In any case, Jamie did mention that there are sufficient channels available at the end-stations for this purposes, but all of these are fast channels. What needs to be decided is if we are going ahead and using the fast channels, or if we need to find slow DAC channels. 
  • I spoke to Koji about gluing the mirrors to the PZTs, and he says we can use superglue, and also to be sure to clean both the mirror and the tip/tilt surfaces before gluing. In any case, all the other hardware issues need to be sorted out first before thinking about gluing the mirrors.

High-Voltage Power Supply

photo_3.JPG

 

Situation at rack 1Y4

 

photo_4.JPG

 Wiring diagram

ASC_schematic.pdf

  8800   Wed Jul 3 21:19:04 2013 gautamConfigurationendtable upgradeplan of action for PZT installation

 This is an update on the situation as far as PZT installation is concerned. I measured the required cable (PZT driver board to PZT) lengths for the X and Y ends as well as the PSL table once again, with the help of a 3m long BNC cable, just to make sure we had the lengths right. The quoted cable lengths include a meter tolerance. The PZTs themselves have cable lengths of 1.5m, though I have assumed that this will be used on the tables themselves. The inventory status is as follows.

  1. Stuff ordered:
    • RG316 LEMO 00 (female) to SMB (female) cables, 10 meters - 6pcs (for the Y-end)
    • RG316 LEMO 00 (female) to SMB (female) cables, 11 meters - 6pcs (for the X-end)
    • RG316 LEMO 00 (female) to SMB (female) cables, 15 meters - 8pcs (6 for the PSL, and two spares)
    • RG316 SMA (male) to open cables, 3 meters - 3pcs (1 each for the X end, Y end and PSL table, for connecting the driver boards to the 100V DC power supply)
    • 10 pin IDC connectors for connecting the DAC interface to the PZT driver boards 
  2. Stuff we have:
    • 40 pin IDC connectors which connect to the DAC interface
    • PZT driver boards
    • PZT mounts
    • Twisted ribbon wire, which will be used to make the custom ribbon to connect the 10 pin IDC to the 40 pin IDC connector

I also did a preliminary check on the driver boards, mainly to check for continuity. Some minor modifications have been made to this board from the schematic shown here (using jumper wires soldered on the top-side of the PCB). I will have to do a more comprehensive check to make sure the board as such is functioning as we expect it to. The plan for this is to first check the board without the high-voltage power supply (using an expansion card to hook it up to a eurocrate). Once it has been verified that the board is getting powered, I will connect the high-voltage supply and a test PZT to the board to do both a check of the board as well as a preliminary calibration of the PZTs.

To this end, I need something to track the spot position as I apply varying voltage to the PZT. QPDs are an option, the alternative being some PSDs I found. The problem with the latter is that the interfaces to the PSD (there are 3) all seem to be damaged (according to the labels on two of them). I tried connecting a PSD to the third interface (OT301 Precision Position Sensing Amplifier), and hooked it up to an oscilloscope. I then shone a laser pointer on the psd, and moved it around a little to see if the signals on the oscilloscope made sense. They didn't on this first try, though this may be because the sensing amplifier is not calibrated. I will try this again. If I can get one of the PSDs to work, mount it on a test optical table and calibrate it. The plan is then to use this PSD to track the position of the reflected beam off a mirror mounted on a PZT (temporarily, using double sided tape) that is driven by feeding small-amplitude signals to the driver board via a function generator. 

 

Misc

The LEMO connector on the PZTs have the part number LEMO.FFS.00, while the male SMB connectors on the board have the part number PE4177 (Pasternack)

Plan of Action:

  • The first task will be to verify that the board is working by the methods outlined above.
  • Once the board has been verified, the next task will be to calibrate a PZT using it. I have to first identify a suitable way of tracking the beam position (QPD or PSD?)
  • I have identified a position in the eurocrate at 1Y4 to install the board, and I have made sure that for this slot, the rear of the eurocrate is not hooked up to the cross-connects. I now need to figure out the exact pin configuration at the DAC interface: the bank is marked 'DAC Channels 9-16' (image attached) but there are 40 pins in the connector, so I need to map these pins to DAC channels, so that when making the custom ribbon, I get the pin-to-pin map right.

DAC_bank.png

 

The wiring scheme has been modified a little, I am uploading an updated one here. In the earlier version, I had mistaken the monitor channels as points from which to log data, while they are really just for debugging. I have also revised the coaxial cable type used (RG316 as opposed to RG174) and the SMB connector (female rather than male).

ASC_schematic.pdf 

 

 

 

 

  4387   Tue Mar 8 15:33:09 2011 kiwamuSummaryGreen Lockingplan on Mar.8th
Today's goal is to measure the contribution from the intensity noise to the beatnote.
 
Plans for today
  - check the ADC for the DCPD that Jenne installed yesterday
  - adjust RF power on the AOM
  - take spectrum of the differential noise and measure the coupling from the intensity noise
  - update the noise budget

Quote: from #4382
This week's goal is to investigate the source of the differential noise and to lower it.

 

  5122   Fri Aug 5 08:08:42 2011 kiwamuSummaryGeneralplan today

Today's main mission is : adjustment of the arm length

 

   + Open the ETMX(Y) door, starting from 9:00 AM

   + Secure the ETMX(Y) test mass by tightening the earthquake stops.

   + Move the ETMX(Y) suspension closer to the door side

   + Inspect the OSEMs and take pictures before and after touching the OSEMs.

   + Level the table

   + Adjust the OSEM positions

   + Move the ETMX(Y) suspension to have designed X(Y)arm length

   + Level the table again

   + Align the ETMX(Y) such that the green beam resonate

  3982   Tue Nov 23 23:13:40 2010 kiwamuSummaryCDSplan: we will install C1LSC

 [Joe, Suresh, Kiwamu]

 We will fully install and run the new C1LSC front end machine tomorrow.

And finally it is going to take care of the IOO PZT mirrors as well as LSC codes. 

 


 (background stroy)

 During the in-vac work today, we tried to energize and adjust the PZT mirrors to their midpoints.

However it turned out that C1ASC, which controls the voltage applying on the PZT mirrors, were not running.

We tried rebooting C1ASC by keying the crate but it didn't come back.

 The error message we got in telnet  was :

   memory init failure !!

 

 We discussed how to control the PZT mirrors from point of view of both short term and long term operation.

We decided to quit using C1ASC and use new C1LSC instead.

A good thing of this action is that, this work will bring the CDS closer to the final configuration. 

 

(things to do)

 - move C1LSC to the proper rack (1X4).

 - pull out the stuff associated with C1ASC from the 1Y3 rack.

 - install an IO chasis to the 1Y3 rack.

- string a fiber from C1LSC to the IO chasis.

- timing cable (?)

- configure C1LSC for Gentoo

- run a simple model to check the health

- build a model for controlling the PZT mirrors

  2568   Wed Feb 3 11:13:15 2010 steveConfigurationGeneralplaned power outage for Sat. Feb 20

The electrical shop has to connect the new power transformer at CES. This means we will have no AC power for ~8 hrs on Saturday, February 20

Is this date good for us to power down ALL equipment in the lab?

Rana:  Yes

  4382   Mon Mar 7 18:20:01 2011 kiwamuSummaryGreen Lockingplans
This week's goal is to investigate the source of the differential noise and to lower it.
 
Plans for tonight
 - realign GREEN_TRANS PD at the PSL table
 - update the noise budget
 - take spectrum of the differential noise
 - investigate a noise coupling to the differential noise especially from the intensity noise
 - update the noise budget again
 
Plans for this week :
 - Auto alignment scripts for green (Kiwamu)
 - connect the end REFL_DC  to an ADC (Kiwamu)
 - make an active phase rotation circuit for the end PDH (undergrads)
 - bounce-roll notches (Suresh)
 - optimization of the suspensions including the input matrices and the Q-values (Jenne)
 - optimization of MFSS (Koji/Rana/Larisa)
 - rewire the mechanical shutter on the 1X9 binary outputs (Steve)

 

  6353   Mon Mar 5 06:11:08 2012 kiwamuSummaryLSCplans

Plans:

  •  DRMI (PRMI) + one arm test before the LVC meeting
  •  Study of the funny sensing matrix and the RAM offset effects before the LVC meeting
  •  Glitch hunting

Action items:

  • MC beam pointing 
    • to make the PZT1 pitch relax
  • OSA setup
    •    a long BNC cable for monitoring the signal in the control room
  • Power budget on the AP table
    • in order to ensure the laser power on each photo diode
  •  POP22/110 sideband monitor
    • installation of an RF amp
    • building a diplexer
    • connect the signals to the demod boards 
  •  Calibration of the demod boards
    • calibrate the conversion loss of the mixers to calibrate all the LSC signals to watts / meter
  •  (1+G) correction for the glitch time series data
  • Simulation study for the RAM offset
    • How much offset do we get due to the RAM ? and how do the offsets screw up the sensing matrix ?
  •  A complete set of the MICH characterization
    •   DC power
    •   Sensing matrix
    •   Noise budget
    •   OSA
    •   Estimation of the RAM offset 
    •  Summarize the results in the wiki
  •  A complete set of the PRMI/DRMI characterization
    •  The same stuff as the MICH characterization
  •  DRMI + one arm test
    •   Monitor the evolution of the sensing matrix during the arm is brought to the resonance

   
 

  1624   Mon May 25 21:31:47 2009 carynUpdatePEMplugged in Guralp channels

Guralp Vert1b and Guralp EW1b are plugged back in to PEM ADCU #10 and #12 respectively. Guralp NS1b remains plugged in. So,  PEM-SEIS_MC1_X,Y,Z should now corrsp to seismometer as before.

  1648   Wed Jun 3 12:31:13 2009 carynUpdatePEMplugged in guralp channels
  5063   Fri Jul 29 18:43:02 2011 Manuel, IshwitaUpdatePEMplugging seismometers to ADC

[Manuel, Ishwita, Jenne, Jamie]

We changed the C1PEM model and the names of the C1:PEM channels.

We reinstalled the blue breakout box, since the purple one still didn't work.

So, now the AA board channels are connected as follows...

C1 = C1:PEM-SEIS_GUR1_X

C2 = C1:PEM-SEIS_GUR1_Y

C3 = C1:PEM-SEIS_GUR1_Z

C4 = C1:PEM-SEIS_GUR2_X

C5 = C1:PEM-SEIS_GUR2_Y

C6 = C1:PEM-SEIS_GUR2_Z

C7 = C1:PEM-SEIS_STS_1_X

C8 = C1:PEM-SEIS_STS_1_Y

C9 = C1:PEM-SEIS_STS_1_Z

C11 = C1:PEM-SEIS_STS_2_X

C12 = C1:PEM-SEIS_STS_2_Y

C13 = C1:PEM-SEIS_STS_2_Z

C14 = C1:PEM-SEIS_STS_3_X

C15 = C1:PEM-SEIS_STS_3_Y

C16 = C1:PEM-SEIS_STS_3_Z

C17 = C1:PEM-ACC_MC1_X

C18 = C1:PEM-ACC_MC1_Y

C19 = C1:PEM-ACC_MC1_Z

C20 = C1:PEM-ACC_MC2_X

C21 = C1:PEM-ACC_MC2_Y

C22 = C1:PEM-ACC_MC2_Z

Although the channels for all 3 STS-2 seismometers are made but only one is installed. So only Channels C1 to C9 are now in operation...

We checked the data from the plugged channels with the Dataviewer. We could see the peak whenever someone jumped in the lab. Even Kiwamu jumped and saw his signal.

  4345   Wed Feb 23 16:34:42 2011 valeraConfiguration pmc lens staged

I put the PMC last mode matching lens (one between the steering mirrors) on a translation stage to facilitate the PMC mode matching.

Currently 4% of incident power is reflected by the PMC. But the reflected beam does not look "very professional" on the camera to Rana - meaning there is too much TEM20 (bulls eye) mode in the reflected beam.

I locked the  PMC  on bulls eye mode and measured  the ratio of the TEM20/TEM00 in transmission to be 1.3%. Thus the PMC mode matching is ~99% and the incident beam HOM content is ~3%.

While working on the PMC I found that the source of PMC "blinking" is not the frequency control signal from MC to the laser (the MC servo was turned off) but possibly some oscillation which could be affected even by a small change of the pump current 2.10 A to 2.08 A. I showed this behaviour to Kiwamu and we decided to leave the the current at 2.08 A for now where things look stable and investigate later.

Attachment 1: PMCrefl.JPG
PMCrefl.JPG
Attachment 2: P1070438.JPG
P1070438.JPG
Attachment 3: P1070439.JPG
P1070439.JPG
  9961   Fri May 16 09:46:05 2014 SteveUpdatePSLpointing monitoring

Quote:

 Tonight I noticed that the drop in PMC transmission was ~1V, more than the usual of ~0.5V from the daily drift.

While re-aligning on the table, I noticed that the misalignment was not from either of the steering mirrors; i.e. I has to walk them both to get the alignment back. This implies that the misalignment is generated far upstream. Maybe the the laser itself is moving. We need some updates from Steve's laser misalignment tracker.

I'd like to replace the paper target with IOO -QPD_POS so we can log it.

  11820   Sat Nov 28 11:46:40 2015 yutaroUpdateLSCpossible error source of loss map measurement

I found that TRY level degraded and the beam shape seen with CCD camera at AS port was splitted when the beam spot on ETMY was not close to the center. This was because dither started not working well. I suspect so because in such a case TRY level went up when I did iteration with TT1 and TT2 after freezing dither. Splitted beam shape indicates that incident light did not match well with the cavity mode.

TRY level for each point was this:

TRYDC
[[ 0.6573      0.8301      0.8983      0.8684      0.6773    ]
 [ 0.7555      0.8904      0.9394      0.8521      0.6779    ]
 [ 0.6844      0.8438      0.9318      0.8834      0.6593    ]
 [ 0.7429      0.8688      0.9254      0.8427      0.6474    ]
 [ 0.7034      0.8447      0.8834      0.8147      0.6966    ]]

 In the worst case, TRY level was 70 % of the maximum level. Assuming that this degrade was totally due to the mode mismatch, this corresponds to ~50 urad difference between the angle of incident light and resonant lighe in the arm (see elog 11819).

  8078   Wed Feb 13 19:09:32 2013 yutaSummaryGeneralpossible explanations to oval REFL beam

[Jenne, Manasa, Jamie, Yuta]

The shape of the REFL beam reflected from PRM is oval after the Faraday.
We tried to fix it by MC spot position centering and by tweaking input TT1/TT2/PRM. But REFL still looks bad (below).

REFL_1044844506.bmp

What has changed since:
  REFL looks OK in mid-Dec 2012. Possibly related things changed are;

  1. New active input TTs with new mirrors installed
  2. Leveling of IMC stack changed a little (although leveling was done after installing TTs)

Possible explanations to oval REFL:
  A. Angled input beam:
    Input beam is angled compared with the Faraday apertures. So, beam coming back from PRM is angled, and clipped by the Faraday aperture at the rejection port.

  B. Mode mis-match to PRM:
    New input TTs have different curvatures compared with before. Input mode matching to PRM is not good and beam reflected from PRM is expanding. So, there's clipping at the Faraday.

  C. Not clipping, but astigmatism:
    New input TTs are not flat. Incident angle to TT2 is ~ 45 deg. So, it is natural to have different tangential/sagittal waist sizes at REFL.

How to check:
  A. Angled input beam:
    Look beam position at the Faraday apertures. If it doesn't look centered, the incident beam may be angled.
   (But MC centering didn't help much......)

  B. Mode mis-match to PRM:
    Calculate how much the beam size will be at the Faraday when the beam is reflected back from PRM. Put some real numbers to curvatures of input TTs for calculation.

  C. Not clipping, but astigmatism:
    Same calculation as B. Let's see if REFL is with in our expectation or not by calculating the ratio of tangential/sagittal waist sizes at REFL.

  8079   Wed Feb 13 19:30:45 2013 KojiSummaryGeneralpossible explanations to oval REFL beam

>> "What has changed since:"

Recently the REFL path has been rearranged after I touched it just before Thanksgiving.
(This entry)

If the lenses on the optical table is way too much tilted, this astigmatism happens.
This is frequently observed as you can find it on the POP path right now.

Also the beam could be off-centered on the lens.

I am not sure the astigmatism is added on the in-air table, but just in case
you should check the table before you put much effort to the in-vacuum work.

  8080   Wed Feb 13 19:41:07 2013 yutaSummaryGeneralpossible explanations to oval REFL beam

We checked that REFL beam is already oval in the vacuum. We also centered in-air optics, including lens, in the REFL path, but REFL still looks bad.

By using IR card in vacuum, PRM reflected beam looks OK at MMTs and at the back face of the Faraday. But the beam looks bad after the output aperture of the Faraday.

  6159   Tue Jan 3 15:49:27 2012 JamieUpdateComputerspossible front-end timing issue

Quote:

Is there a reason the framebuilder status light is red for all the front ends?

Also, I reenabled PRM watchdog.

Apparently there is a bug in the timing cards having to do with the new year roll-over that is causing front-end problems.  From Rolf:

For systems using the Spectracom IRIG-B cards for timing information, the code did not properly roll over the time for
2012 (still thinks it is 2011 and get reports from DAQ of timing errors (0x4000)). I have made a temporary fix for this
in the controller.c code in branch-2.3, branch-2.4 and release 2.3.1. 

I was going to check to see if the 40m is suffering from this. I'll be over to see if that's the problem.

  6168   Wed Jan 4 09:06:50 2012 steveUpdateComputerspossible front-end timing issue

Quote:

Quote:

Is there a reason the framebuilder status light is red for all the front ends?

Also, I reenabled PRM watchdog.

Apparently there is a bug in the timing cards having to do with the new year roll-over that is causing front-end problems.  From Rolf:

For systems using the Spectracom IRIG-B cards for timing information, the code did not properly roll over the time for
2012 (still thinks it is 2011 and get reports from DAQ of timing errors (0x4000)). I have made a temporary fix for this
in the controller.c code in branch-2.3, branch-2.4 and release 2.3.1. 

I was going to check to see if the 40m is suffering from this. I'll be over to see if that's the problem.

 The problem is the same as yesterday.

Attachment 1: rtntstat.png
rtntstat.png
  6574   Thu Apr 26 18:15:59 2012 JamieUpdateCDSpossible issue with mx_stream on front ends

I'm noticing what appears to be occasional failures of mx_stream on the front end machines.  It doesn't happen that frequently, but I've noticed it a couple of times already since the upgrade.

The symptom is that the DC Status goes to "0xbad" (red) and the "FE NET" goes red for all models on a given front end.

The solution seems to be restarting mx_stream on the given front end:    sudo  /etc/init.d/mx_stream restart"

There is nothing in the mx_stream log:

 controls@c1sus ~ 0$ cat /opt/rtcds/caltech/c1/target/fb/mx_stream_logs/c1sus.log 
 c1x02
 c1sus
 c1mcs
 c1rfm
 c1pem
 mmapped address is 0x7f43740ec000
 mapped at 0x7f43740ec000
 mmapped address is 0x7f43700ec000
 mapped at 0x7f43700ec000
 mmapped address is 0x7f436c0ec000
 mapped at 0x7f436c0ec000
 mmapped address is 0x7f43680ec000
 mapped at 0x7f43680ec000
 mmapped address is 0x7f43640ec000
 mapped at 0x7f43640ec000
 send len = 263596
 Connection Made

but I do see some funny messages in the front end dmesg:

 [200341.317912] DXH Adapter 0 : Heartbeat alive-check for node=12 failed (cnt=8387 state=0x1 deb=0 val=0).
 [200341.318670] DXH Adapter 0 : Session for node 12 is disabled - Status = 0x5
 [200341.319062] Session callback reason=1 status=5 target_node=12
 [200341.319069] Session callback reason=3 status=0 target_node=12
 [200341.359534] (map_table_check_access:752):my id 1 ->  remote id 2 : entry was valid - is now tentatively valid
 [200341.859584] DXH Adapter 0 : Probe failure for node=12 - disabling session probeStatus=0x40000f02
 [200341.860335] DXH Adapter 0 : Session for node 12 is disabled - Status = 0x3
 [200341.860728] Session callback reason=1 status=3 target_node=12
 [200374.006111] DXH Adapter 0 : Set reachable remote node list.
 [200409.020670] DXH Adapter 0 : Set reachable remote node list.
 [200409.021076] DXH Adapter 0 : Session for node 12 is deleted - Status = 0x0
 [200409.021468] Session callback reason=5 status=0 target_node=12
 [200412.362824] (map_table_insert:648):** successfully inserted **(valid unicast) inst 0 node 1->0 fwd 0 fwd_tp 4 egress 0
 [200418.025994] (map_table_check_access:752):my id 1 ->  remote id 0 : entry was valid - is now invalid
 [200418.025998] (map_table_insert:648):** successfully inserted **(valid unicast) inst 0 node 1->2 fwd 0 fwd_tp 4 egress 0
 [200421.743916] Session callback reason=0 status=0 target_node=12
 [200422.073776] DXH Adapter 0 : Set reachable remote node list.
 [200422.342446] Session callback reason=7 status=0 target_node=12
 [200422.342454] DXH Adapter 0 : Session for node 12 is ok.

I'm awaiting feedback from experts.

 

  12096   Thu Apr 28 08:49:47 2016 SteveUpdatePEMpossible noise sources schedule

Building:         Campus (see attached Map)        

       

Date:              Manhole 1 - May 3 through May 5

Manhole 2 – May 6 through May 10

 Manhole 2 - May 16 through May 19

Manhole 3 – May 11 through May 19           

          

Time:              Noise:  7:00 a.m. To 5:00 p.m.

                  Access: 24 Hours a day

           

Interruption:      Noise/Vehicular & Pedestrian Access

                  Storm Drain Manholes

         

*In order to repair 3 manholes associated with a large storm drain that runs north-south through the campus, work will take place at the

3 manholes shown on the map. This work will interrupt vehicular and pedestrian access on the paths adjacent to the manholes. Though the work at Manholes 1 and 2 will allow vehicular and pedestrian access around the manholes, the work at Manhole 3 will completely block the driveway running south from the southeast corner of Parking Lot 11. Noise will also be created by the repair

 

 

Attachment 1: Campus_B&W_Map-2.pdf
Campus_B&W_Map-2.pdf
  109   Thu Nov 15 18:37:06 2007 tobinUpdateComputerspossible replacement for linux1's disk
It looks like the existing disk in linux1 is a Seagate ST380013A (this can be found either via the smartctl utility or by looking at the file /proc/ide/hda/model). It appears that you can still buy this disk from amazon, though I think just about any ATA disk would work. I'll ask Steve to buy one for us.
  13149   Fri Jul 28 20:22:41 2017 JamieUpdateCDSpossible stable daqd configuration with separate DC and FW

This week Jonathan Hanks and I have been trying to diagnose why the daqd has been unstable in the configuration used by the 40m, with data concentrator (dc) and frame writer (fw) in the same process (referred to generically as 'fb').  Jonathan has been digging into the core dumps and source to try to figure out what's going on, but he hasn't come up with anything concrete yet.

As an alternative, we've started experimenting with a daqd configuration with the dc and fw components running in separate processes, with communication over the local loopback interface.  The separate dc/fw process model more closely matches the configuration at the sites, although the sites put dc and fwprocesses on different physical machines.  Our experimentation thus far seems to indicate that this configuration is stable, although we haven't yet tested it with the full configuration, which is what I'm attempting to do now.

Unfortunately I'm having trouble with the mx_stream communication between the front ends and the dc process.  The dc does not appear to be receiving the streams from the front ends and is producing a '0xbad' status message for each.  I'm investigating.

  11757   Thu Nov 12 10:22:33 2015 SteveUpdatePEMpossible vibration for 4days

Building:         San Pasqual walkway East to West

                  (Between Holliston & Wilson)         

 

Date:             Thursday 11-12-15 to Wednesday 11-18-15

 

Time:             Between 6:00 a.m. and 4:00 p.m. each day        

 

Notification:     Possible Noise Vibration

 

Contact:          Ken Lewis (626) 298-2037       

 

* Plumbing contractor will be inspecting and water jetting Storm drains

Type of interruption: (Some vehicle noise and small vibrations limited to close surrounding area)

Areas effected: San Pasqual walkway from Holliston Street to Wilson)

Potential effects: storm drain loss of use

Reason for interruption: Storm drain cleaning in preparation for rainy season

 

  13310   Mon Sep 11 23:31:50 2017 johannesUpdateCameraspost-vent camera capture comparison

The latest pre-unintended vent captures of the test mass face cameras were taken on June 2nd, 2017. Only exposures for ITMYF, ETMYF, and ETMXF exist in /users/sensoray/SensorayCaptures/. I took new captures for those three after locking the arms and having the dither-alignment on for 5+ minutes (exposures were taken after turning the dithering off). The capture script is choking on ITMXF, saying the channel can't lock on. Maybe that's why there's also no reference image for it. Capturing QUAD3, which shows ITMXF in the lower right corner, works, but we don't have a capture for reference. I also recorded dark fields after closing the PSL shutter. Naturally, these don't subtract out as well for the three-month old pictures, but it's actually not terrible and qualitatively one can still compare the subtracted images

Visually, ITMYF and ETMYF do not show a dramatic difference between then and now. ETMXF however, does. To get a numerical estimate for the difference in counts, I worked with the subtracted images and placed an aperture about 1.5x the size of the visible beam blob. I summed up the pixel values inside and subtracted the sum of the pixel values of an equally sized area from the upper left corner of the respective image, which looks free of subtraction artifacts and looks qualitatively similar to the background in the central region.

The pixel sum has gone up by about 50% between the exposures. I still have to do the same for the YARM optics but don't expect such a large discrepancy. Unfortunately we're missing those ITMYF expsures...

All pictures are organized in this format:

Pre-vent exposure Post-vent exposure
Pre-vent subtracted Post-vent subtracted

 

ITMYF

   

   

ETMYF

   

   

ETMXF

   

   

Attachment 11: ETMXF_pre_sub.bmp
  13334   Tue Sep 26 22:11:08 2017 johannesUpdateCameraspost-vent camera capture comparison

I configured the remaining GigE-Camera to work on the 40m network. We currently have 3 operational Basler cameras:

The 120gm's have been assigned the IPs 192.168.113.152  (was already configured) and 192.168.113.153 (freshly configured) and have been labeled accordingly. Note that it was not necessary to connect the out-of-the-box camera directly to a dedicated ethernet adapter whose IP was set manually to 169.254.0.XXX as pointed out in earlier posts - a few seconds after connecting the camera to the control room switch (with PoE adapter to power it) the camera showed up in the configuration software tool which is launched via

/opt/rtcds/caltech/c1/scripts/GigE/pylon5/bin/./IpConfigurator

and can be assigned a corrected, static IP.

We have a plethora of 2" tubes for the lens assembly, but not a great variety of focal lengths for 2" lenses. Present with the camera gear were two f=250 mm and one f=150 mm 2" lenses with a NIR broadband AR coating

To determine the lens positions relativ to the sensor I assumed that the camera we're setting up looks at its test mass from a distance of 1m. Using the two available focal lengths we can look for solutions which have reasonable lens separations <~10cm and suitable magnification. We primarily want to image the central mirror area onto a 1/4" sized sensor, which can be achieved with a magnification of ~1/8.

I chose a lens separation of 6cm, which gives a theoretical magnification of -.12 and a sensor-lens 2 distance of 7.95 cm. I placed the lenses accordingly in the tubes and checked the focusing with Gautam's help:

       

It's pretty close to what we would expect. We will do the calibration using the auxiliary laser on the PSL table. For this I temporarily routed a fiber from the PSL enclosure to the SP table. Since the main cable hole is sort of cramped it's going in through a gap near the ceiling instead.  

 

Attachment 1: lens_distance.pdf
lens_distance.pdf
  15550   Sun Aug 30 11:29:33 2020 ranaUpdateGeneralpower blink?

My power at home winked out for a second this morning, but it looks like either nothing happened in the 40m lab or else it rode it out.

MC is locked - lost lock around 11:25 AM and then relocked.

  4448   Mon Mar 28 16:24:35 2011 kiwamuUpdateGreen Lockingpower budget on PSL table

   I measured some laser powers associated with the beat-note detection system on the PSL table.

The diagram below is a summary of the measurement. All the data were taken by the Newport power meter.

 The reflection from the beat-note PD is indeed significant as we have seen.

In addition to it the BS has a funny R/T ratio maybe because we are using an unknown BS from the Drever cabinet. I will replace it by a right BS.

RFPD.png

(background)

 During my work for making a noise budget I noticed that we haven't carefully characterize the beat-note detection system.

The final goal of this work is to draw noise curves for all the possible noise sources in one plot.

To draw the shot noise as well as the PD dark noise in the plot, I started collecting the data associated with the beat-note detection system.

 

(Next actions)

 * Estimation and measurement of the shot noise

 * measurement of the PD electrical noise (dark noise)

 * modeling for the PD electrical noise

 * measurement of the doubling efficiency

 * measurement of an amplitude noise coupling in the frequency discriminators

  6355   Mon Mar 5 14:10:35 2012 kiwamuUpdateLSCpower budget on the AP table

I checked the laser powers on the AP table and confirmed that their powers are low enough at all the REFL photo diodes.

When the HWP( which is for attenuating the laser power with a PBS) is at 282.9 deg all of the REFL diodes receives about 5 mW.

This will be the nominal condition. 

If the HWP is rotated to a point in which the maximum laser power goes through, the diodes get about 10 mW, which is still below the power rate of 18 mW (#6339).

I used the Coherent power meter for all the measurements.

The table below summarizes the laser powers on the REFL diodes and the OSA. Also the same values were noted on the attached picture.

 

 nominal power [mW]

(when HWP is at 282.9 deg)

expected max power [mW]

(when HWP is at a point where the max power goes through)

REFL11 5.5 10
REFL33 4.5 10
REFL55 5.3 10
REFL165 4.8 10
REFL OSA 0.7 0.7

 

A note:
I found that the OSA for the REFL beam was receiving a unnecessary bright laser.
So I put an ND1 attenuator stacked on the existing ND2 attenuator. The laser power entering in the OSA is currently at 0.7 mW.
Attachment 1: power_budget.png
power_budget.png
  12593   Thu Nov 3 08:07:52 2016 SteveUpdateGeneralpower glitch

Building:         Campus Wide         

       

Date:             Thursday 11/03/16 at Approx. 6:20 a.m.   

          

Notification:     Unplanned City Wide Power Glitch Affecting Campus   

 

*This is to notify you that the Caltech Campus experienced a campus wide power glitch at approx. 6:20 a.m. this morning.

The city was contacted and they do not expect any further interruptions related to this event.

 

The vacuum was not effected. ITM sus damping restored. IFO room air conditions on.

PSL Innolight and ETMY Lightwave lasers turned on

 

Attachment 1: powerGlitch.png
powerGlitch.png
  12696   Mon Jan 9 09:18:47 2017 SteveUpdatePEMpower glitch

There was a power glitch last night around 1:15am

The vacuum was not effected.

PSL laser turned on, PMC locked, PSL shutter opened and MC locked.

IR lasers at the ends turned on.

East arm air cond turned on.

The computers are all done.

The last power glitch was at Nov 3, 2016

 

 

Attachment 1: MondayMorning.png
MondayMorning.png
  12700   Tue Jan 10 21:47:00 2017 ranaUpdateCDSpower glitch

Does "done" mean they are OK or they are somehow damaged? Do you mean the workstations or the front end machines?

The computers are all done.

megatron and optimus are not responding to ping commands or ssh -- please power them up if they are off; we need them to get data remotely

  12594   Thu Nov 3 11:33:24 2016 gautamUpdateGeneralpower glitch - recovery

I did the following:

  • Hard reboots for fb, megatron, and all the frontends, in that order
  • Checked time on all FEs, ran sudo ntpdate -b -s -u pool.ntp.org where necessary
  • Restarted all realtime models
  • Restarted monit on all FEs
  • Reset Marconi to nominal settings, fCarrier=11.066209MHz, +13dBm amplitude
  • In the control room, restarted the projector and set up the usual StripTool traces
  • Realigned PMC
  • Slow machines did not need any touchups - interestingly, ITMX did not get stuck during this power glitch!

There was a regular beat coming from the speakers. After muting all the channels on the mixer and pulling the 3.5mm cable out, the sound persisted. It now looks like the mixer is broken sad

     ProFX8v2

 

  12702   Wed Jan 11 16:35:03 2017 gautamUpdateCDSpower glitch - recovery progress

[lydia, ericq, gautam]

We set about following the instructions linked in the previous elog. A few notes/remarks:

  1. It is important to run the ntpdate commands before restarting the models. Sometimes, multiple restarts of the models were required to turn all the indicator blocks on the MEDM screen green.
  2. There was also an issue of multiple ntpd processes running on the same machine, which obviously caused all sorts of timing havoc. EricQ helped us diagnose and fix these. At the moment, all the lights are green on the CDS status MEDM screen
  3. On the hardware side, apart from the usual suspects of frontends/megatron/optimus/fb needing to be rebooted, I noticed that the ETMX OSEM lights were off on the control room monitors. Investigation pointed to the 2 20V sorensens at the X end outputting 0V, 0A after the power glitch. We turned down both dials, and then gradually ramped them up again. Both Sorensens now read +/-20V, 0.3A, which is in agreement with the label stuck onto them.
  4. Restarted MC autolocker and FSS Slow scripts on megatron. I have not yet looked at the status of the nds2 server on megatron.
  5. 11 MHz Marconi has yet to be restarted - but I am unable to get even the IMC locked at the moment. For some reason, the RMS of the MC1 and MC3 coils are way higher than what I am used to seeing (~5mV rms as compared to the <1mV rms I am used to seeing for a damped optic). I will investigate further. Leaving MC autolocker disabled for now.
  12701   Tue Jan 10 22:55:43 2017 gautamUpdateCDSpower glitch - recovery steps

Here is a link to an elog with the steps I had to follow the last time there was a similar power glitch.

The RAID array restart was also done not too long ago, we should also do a data consistency check as detailed here, if not already..

If someone hasn't found the time to do this, I can take care of it tomorrow afternoon after I am back.

Quote:

Does "done" mean they are OK or they are somehow damaged? Do you mean the workstations or the front end machines?

The computers are all done.

megatron and optimus are not responding to ping commands or ssh -- please power them up if they are off; we need them to get data remotely

 

  12699   Tue Jan 10 16:20:11 2017 SteveUpdateCDSpower glitch......Raid is rebuilding

Jamie started the fm40m Raid rebuilding. It has been beeping since the power outage.

Summary pages have no reading since power glitch.

 

Attachment 1: rebuilding_in_progress.png
rebuilding_in_progress.png
  5270   Fri Aug 19 15:31:53 2011 steveUpdateGeneralpower interruption rescheduled to 10-1-2011

                UTILITY & SERVICE INTERRUPTION

**PLEASE POST**

 

Building:               Central Engineering Services (C.E.S.)

          LIGO Gravitational Physics building adjacent to C.E.S. 40M- Lab

          Safety Storage adjacent to CES

          Steele House 

          Keck Lab

 

Date:                   Saturday, October 1, 2011

Time:                   8:00 a.m. To 9:00 a.m.            

Interruption:   Electricity

Contact:                Mike Anchondo ext. 4999  Tom Brennan 4984

*This interruption is required for maintenance of high voltage switchgear in Campus Sub Station.

(If there is a problem with this Interruption, please notify

 the Service Center X-4717 or the above Contact as soon as possible.

 If no response is received we will proceed with the interruption.)

         

                                Reza Ohadi,

                                Director, Campus Operations & Maintenance

  12808   Tue Feb 7 16:23:49 2017 SteveUpdateGeneralpower interruption tomorrow

                                                                                                                                   received this note: at 4:11pm Tuesday, Feb 7, 2017

**PLEASE POST**

 

Building:         Campus

    

Date:             Wednesday, February 8, 2017

          

Time:             7:30 AM – 8:30 AM  

 

Contact:          Rick Rodriguez x-2576

           

Pasadena Water and Power (PWP) will be performing a switching operation of the

Caltech Electrical Distribution System that is expected to be transparent to Caltech,

but could result in a minor power anomaly that might affect very sensitive equipment.

 

IMPACT: Negligible impact......?

There may be temporary  power interruption tomorrow!

PS:we did not see any effect   

ELOG V3.1.3-