40m QIL Cryo_Lab CTN SUS_Lab TCS_Lab OMC_Lab CRIME_Lab FEA ENG_Labs OptContFac Mariner WBEEShop
  40m Log, Page 231 of 344  Not logged in ELOG logo
ID Date Authorup Type Category Subject
  15342   Thu May 21 15:31:26 2020 gautamUpdateComputer Scripts / ProgramsNDS2 service restarted

The service had failed at 16:09 yesterday. I just restarted it and am now able to fetch data again. 

Unrelated to this work: I restarted the httpd service on nodus a couple of times this afternoon while experimenting with the summary pages.

Quote:

Please try it out and tell me about any problems in getting fresh data.

  15343   Fri May 22 01:43:18 2020 gautamUpdateElectronicsRF electronics trouble

To test a hypothesis, I have left the PSL shutter closed. I notice significant glitches in the dark electronics offsets on all the 11 MHz photodiode I/Q demodulated input channels, which appear coherent. These are non-negligible in magnitude - for now they are uncalibrated in cts, but for an estimate, the POX11 channel shows a shift of ~20 cts (~200uV at the input to the whitening board), while the PDH fringe is ~200 cts pk2pk. A first look is in Attachment #1. The fact that it's in all the 11 MHz channels makes me suspect something in the RF chain, maybe some amplifier? I'll open the shutter tomorrow.

  15347   Tue May 26 01:58:57 2020 gautamUpdateElectronicsSome electronics thoughts

A big factor in how much IFO locking activities can take place is how cooperative the IMC is.

Since the c1psl upgrade, the IMC duty cycle has definitely deteriorated. I took a measurement of the dark noise at the IMC error point with 1 Hz FFT binwidth, with all electrical connections to the IMC servo board except the Acromag and Eurocrate power disconnected. I was horrified at the prominence of 60 Hz harmonics - see Attachment #1. In the past, this kind of feature has been indicative of some error in the measurement technique - but I confirmed that the lines remain even if I unplug the GPIB box, and all combinations of floating/grounded inputs that I tried. We know for sure that there is some excess noise imprinted on the laser light post upgrade. While these lines almost certainly are not responsible for the PCdrive RMS going bonkers, surely this kind of electrical situation isn't good?

Attachment #2 shows the same information translated to frequency noise units, taking into account the complementary sensitivity function, L/(1+L) - the sum contribution of the 60 Hz peaks to the RMS is ~11.5% of the total over the entire band (c.f. 1.7 % that is expected if the noise at multiples of 60 Hz was approximately equal to the surrounding noise levels). Moreover, the measured RMS is 55 times higher than a LISO model. 

How can this be fixed?

  15348   Tue May 26 02:15:36 2020 gautamUpdateLSCLock acquisition portal entry

Summary:

Provided the IMC is cooperative, the input pointing isn't drifting, and the RF offsets aren't jumping around too much, the locking sequence is now pretty robust.

Details:

Most of the analysis uses data between the GPS times 1274418176 and 1274419654 that are recorded to frames.

  15349   Tue May 26 02:31:00 2020 gautamUpdateLSCLock acquisition sequence

Here, I provide some details of the sequence. Obviously, I am presenting one of the quickest transitions to the fully locked state, I don't claim that every attempt is so smooth. But it is pretty cool that the whole thing can be done in ~3 minutes.

See Attachment #1 for the labels.

  • A --- Arms are locked on POX/POY, and EX/EY lasers are also locked to their respective arms. The phase tracker outputs are averaged in preparation for transitioning control from POX/POY to ALS. 
  • B --- Aforementioned transition has been realized. CARM offset of -4 is applied. Based on this calibration, this is ~ 4 nm.
  • C --- PRM is aligned in preparation for 3f vertex locking. Between C and D, the long pause is because I also use this time to DC couple the ITM Oplev servos, which requires some averaging. 
  • D --- PRMI is locked. CARM offset reduction begins. Between D and E, I scan CARM through a resonance, and look at the necessary offset requried in the CARM_B (=RF) path. It is a mystery to me why this is required.
  • E --- Ramp CARM offset completely to 0. Twiddle CARM_A and DARM_A offsets (=ALS path) to maximize the arm transmitted powers. Between E and F, you can see that the arm powers stabilize somewhat before any RF control is engaged (more on this later), which means we are approximately in the linear regime of the CARM PDH signal, and the switchover can be effected. As I write this, I wonder if there is any benefit to normalizing the REFL_11 error signal (=CM_SLOW) by the arm transmission for a broader capture range?
  • F --- CARM_B and DARM_B (=RF) paths engaged. I ramp off the ALS servos between F and G using a 10 second ramptime.
  • G --- IFO is now under RF control, ALS control has been turned off completely.
  • H --- Rudimentary ASC is enabled. The ITMs are already running with DC coupled Oplev servos, and for the ETMs, I use the Transmon QPDs. The loop shapes/gains for this part haven't been finalized yet, but some improvement in the stability is seen.

This particular lock held for ~20 minutes so I could run some loop characterization measurements etc.

I am struggling to explain:

  1. Why POP22 goes to 0 when we zero the CARM offset? The arm length is such that the 2f fields don't experience any abrupt changes in reflectivity from the arm cavity for a wide range of offsets. This signal is the trigger signal for the PRMI LSC control - right now, I get around this problem by mixing in some amount of POP DC once the PRMI is locked. But if the lock is lost, this requires some EPICS button gynmastics to try and salvage the lock... I guess the 1f field components experience a different phase on reflection at various offsets, so maybe I should be looking at sqrt(POP22_I^2 + POP22_Q^2) instead of just POP22_I.
  2. Why is an error point offset required in the CARM RF path?
  15350   Tue May 26 02:37:19 2020 gautamUpdateLSCDARM loop measurement and fitting

Summary:

In order to estimate the free-running DARM displacement noise, I measured the DARM OLTF using the usual IN1/IN2 prescription. The measured data was then used to fit some model paramters for a loop model that can be used over a larger frequency range.

Details:

  • Attachment #1 shows an overlay of the measured and modelled TFs.
  • Attachment #2 shows the various components that went into building up this model. 
    • The digital AA and AI filter coefficients were taken from the RTCDS code.
    • The analog AA and AI filter zpks were taken from here and here respectively.
    • CDS filters taken from the banks enabled. The 20Hz : 0Hz z:p filter in the CARM_B path is also accounted for, as have the violin-mode notches.
    • Pendulum TF is just 1/f^2, the overall scaling is unimportant because it will be fitted (in combination with the overall scaling uncertainty on the DC optical gain), but I used a value of 10 nm/f^2 which should be in the right ballbark.
    • The optical gain includes the DARM pole at ~4.5 kHz for this config.
  • With all these components, to make the measurement and fit line up, I added two free parameters - an overall gain, and a delay. 
    • NLSQ minimizer was used to find the best-fit values for these parameters.
    • I'm not sure what to make of the relatively large disagreement between measurement and model below 100 Hz - I'm pretty sure I got all the CDS filters included...
    • Moreover, I don't have a good explanation for why the best-fit delay is 400 us. One RTCDS clock cycle is onyl 60 us, and even with an extra clock cycle for the RFM transfer, I still can't get up to such a high delay...

In summary, the UGF is ~150 Hz and phase margin is ~30 deg. This loop would probably benefit from some low-pass filter being turned on.

  15351   Tue May 26 03:01:35 2020 gautamUpdateLSCCARM loop

Summary:

I am able to realize ~8 kHz UGF with ~60 degrees of phase margin on the CARM loop OLTF (combination of analog and digital signal paths).

Details:

  • Attachment #1 shows the measured OLTF.
  • The measurement is made by using the "EXC A" bank on the CM board, with an SR785. With this technique, the measurement will be poor where the loop gain is high, as the excitation will be squished. Nevertheless, we can estimate the behavior in those regimes by using a model, and fitting it to the regions where the measurement is valid (in this case, above ~1 kHz).
  • This measurement was made with IN1 Gain = +4 dB, AO gain = 0 dB, and IMC IN2 gain = 0 dB.
  • The regular boost has been enabled, but no super-boosts yet, mainly because I think they consume too much phase close to the UGF. 
  • The modeling/fitting of this, including a more thorough characterization of the crossover, will follow...
  15352   Tue May 26 03:06:59 2020 gautamUpdateLSCPRFPMI sensing matrix

Summary:

The response of the PRFPMI length degrees of freedom as measured in the LSC PDs was characterized. Two visualizations are in Attachment #1 and Attachment #2.

Details:

  • The sensing matrix infrastructure in the c1cal model was used.
  • The oscillator frequencies are set between 300 - 315 Hz.
  • Notch filters at these frequencies were enabled in the CDS filter banks, to prevent actuation at these frequencies (except for CARM, in which case the loop gain is still non-negligible at ~300 Hz, this correction has not yet been applied).
  • Mainly, I wanted to know what the DARM sensing response in AS55_Q is. 
    • The measurement yields 2.3e13 cts/m. This is a number that will be used in the noise budget to convert the measured DARM spectrum to units of m/rtHz.
    • We have to multiply this by 10/2^15 V/ct, undo the 6dB whitening gain on the AS55_Q channel, and undo the ~5x gain from V_RF to V_IF (see Attachment #4 of this), to get ~0.69 GV/m from the RFPD.
    • The RF transimpedance of AS55_Q is ~550 ohms, and accounting for the InGaAs responsivity, I get an optical gain of 1.8 MW/m. Need to check how this lines up with expectations from the light levels, but seems reasonable.
    • Note that T_SRM is 10%, we dump 70% of the output field into the unused OMC, and there is a 50/50 BS splitting the light between AS55 and AS110 PDs. Assuming 90% throughput from the rest of the chain, we are only sensing ~1.3 % of the output DARM field.
  • Apart from this, I can also infer what the matrix elements / gains need to be for transitioning the PRMI control from 3f to 1f signals. To be done...
  • I found these histograms in Attachment #2 to be a cute way of (i) visualizing the variance in the magnitude of the sensing element and (ii) visualizing the separation between the quadratures, which tells us if the (digital) demod phase needs to be modified.
    • The sensing lines were on for 5 minutes (=300 seconds) and the FFT segment length is 5 seconds, so these histograms are binning the 60 different values obtained for the value of the sensing element.
    • The black dashed lines are "kernel density estimates" of the underlying PDFs
    • I haven't done any rigorous statistical analysis on the appropriateness of using this technique for error estimation, so for now, they are just lines...
  15353   Tue May 26 03:26:58 2020 gautamUpdateLSCPreliminary noise budget

Summary:

This isn't meant to be a serious budget, mainly it was to force myself to write the code for generating this more easily in the future.

Details:

  • DARM OLTF model from here was used to undo the loop to convert the in-loop measurement to a free-running estimate.
  • The AS55 PD channels were whitened to reduce the effect of ADC noise.
  • To measured channel was 'C1:LSC-DARM_IN1_DQ'.
    • Some care needs to be taken when applying the conversion from counts to meters using the sensing element measured here.
    • This is because the sensing matrix measurement was made using the response in the channel 'C1:LSC-AS55_Q_ERR_DQ'.
    • Between 'C1:LSC-DARM_IN1_DQ' and 'C1:LSC-AS55_Q_ERR_DQ' there is a scalar gain of 1e-4, and a z:p = 20:0 filter.
    • These have to be corrected for when undoing the loop, since the measurement point is 'C1:LSC-DARM_IN1_DQ'. 
  • The "Dark noise" trace was measured with the PSL shutter closed, but all CDS filters up to 'C1:LSC-DARM_IN1_DQ' enabled as they were when the DARM measurement was taken.
  • It would be interesting to see what the budget looks like once the DARM loop gain has been turned down a bit, some low-pass filtering is enabled, and the vertex DoFs are transitioned to 1f control which is hopefully lower noise.
  15355   Tue May 26 14:32:44 2020 gautamUpdateLSCArm transmission RIN

Summary:

The measured RIN of the arm cavity transmission when the PRFPMI is locked is ~10x in RMS relative to the single arm POX/POY lock. It is not yet clear to me where the excess is coming from.

Details:

Attachment #1 shows the comparison.

  • For the PRFPMI lock, the ITM Oplev Servos are DC coupled, and the ETM QPD ASC servos are also enabled.
  • Admittedly, the PD used in the POX/POY lock case is the Thorlabs PD while when the PRFPMI is locked, it is the QPD.
  • I found that there isn't really a big difference in the RIN if we normalize by the IMC transmission or not (this is what the "un-normalized" in the plot legend is referring to).  A scatter plot of TRX vs TRY and TRX/MCtrans vs TRY/MCtrans have nearly identical principal components. 
  • To convert to RIN, I divided the ocmputed spectra by the mean value of the data stream. For the POX/POY lock, the arm transmission is normalized to 1, so no further manipulation is required.
  • The spectra are truncated to 512 Hz because the IMC sum channel is DQ-ed at 1 kHz, but because of the above bullet point, in principle, I could calculate this out to higher frequencies.
  15356   Tue May 26 16:00:06 2020 gautamUpdateLSCPower buildup diagnostics

Summary:

I looked at some DC signals for the buildup of the carrier and sideband fields in various places. The results are shown in Attachments #1 and #2.

Details:

  • A previous study may be found here.
  • For the carrier field, REFL, POP and TRX/TRY all show the expected behavior. In particular, the REFL/TRX variation is consistent with the study linked in the previous bullet.
  • There seems to be some offset between TRX and TRY - I don't yet know if this is real or just some PD gain imbalance issue.
  • The 1-sigma variation in TRX/TRY seen here is consistent with the RMS RIN of 0.1 evaluated here.
  • For the sideband powers, I guess the phasing of the POP22 and AS110 photodiodes should be adjusted? These are proxies for the buildup of the 11 MHz and 55 MHz sidebands in the vertex region, and so shouldn't depend on the arm offset, and so adjusting the digital demod phases shouldn't affect the LSC triggering for the PRMI locking, I think.
  • Based on this data, the recycling gain for the carrier is ~12 +/- 2, so still undercoupled. In fact, at some points, I saw the transmitted power exceed 300, which would be a recycling gain of ~17, which is then nearly the point of critical coupling. REFLDC doesn't hit 0 because of the mode mismatch I guess.
  15361   Thu May 28 18:36:45 2020 gautamUpdateLSCArm transmission RIN

I agree, I think the PRC excess angular motion, PIT in particular, is a dominant contributor to the RIN. Attachments #1-#3 support this hypothesis. In these plots, "XARM" should really read "COMM" and "YARM" should really read "DIFF", because the error signals from the two end QPDs are mixed to generate the PIT and YAW error signals for these ASC servos - this is some channel renaming that will have to be done on the ASC model. The fact that the scatter plot between these DoFs has some ellipticity probably means the basis transformation isn't exactly right, because if they were truly orthogonal, we would expect them to be uncorrelated?

  • In the corner plots, I am plotting the error signals of the ASC servos and the arm transmission. POP feedback is not engaged, but some feedback control to the ETMs based on the QPD signals is engaged.
  • In the coherence plot, I show the coherence of the ASC error signals with the POP and TR QPD based error signals, under the same conditions. The coherence is high out to ~20 Hz.

I guess what this means is that the stability of the lock could be improved by turning on some POP QPD based feedback control, I'll give it a shot.

Quote:

- PRC TT misalignment (~3Hz)

Don't can you check the correlation between the POP QPD and the arm RIN

  15364   Wed Jun 3 01:34:53 2020 gautamUpdateLSCLock acquisition update portal

Highlights:

  • With better ASC servos implemented, I think the lock stability has been improved. 
  • Arm transmission of ~350 was stably maintained (PRG~20, overcoupled). It went as high as 410, so this is now very close to the highest (~425) I've ever managed to get.
  • I was trying to get the vertex transitioned to 1f control but it remains out of reach for now.  The noise at ~100 Hz is dominated by MICH-->DARM coupling (as judged by coherence, I haven't yet done the broadband noise injection characterization). I figured I'd try the 1f transition before thinking about feedforward.
  • The biggest problems remain flaky electronics (poor IMC duty cycle, jumping RF offsets, newly glitchy seismometer, ...)

Details:

  15365   Wed Jun 3 01:40:13 2020 gautamUpdateElectronicsMore electronics woes

There were many locklosses from the point where the arm powers were somewhat stabilized. Attachments #1 and #2 show two individual locklosses. I think what is happening here is that the BS seismometer X channel is glitching, and creating a transient in the angular feedforward filter that blows the lock. The POP QPD based feedback loop cannot suppress this transient, apparently. For now, I get around this problem by boosting the POP QPD feedback loop a bit, and then turning the feedforward filters off. The fact that the other seismometer channels don't report any transient makes me think the problem is either with the seismometer itself, or the readout electronics. The seismometer masses were recently recentered, so I'm leaning towards the latter.

I didn't explicitly check the data, but I am reasonably certain the same effect is responsible for many PRMI locklosses even with the arms held off resonance (though the tolerance to excursions there is higher). Pity really, the feedforward filters were a big help in the lock acquisition...

  15366   Wed Jun 3 01:46:14 2020 gautamUpdateLSCCARM loop

Summary:

The CARM loop now has a UGF of ~12 kHz with a phase margin of ~60 degrees. These values of conventional stability indicators are good. The CARM optical gain that best fits the measurements is 9 MW/m.

I've been working on understanding the loop better, here are the notes.

Details:

Attachment #1 shows a block diagram of the loop topology.

  • The "crossover" measurement made at the digital CARM error point, and the OLG measurement at the CM board error point are shown.
  • I've tried to include all the pieces in the loop, and yet, I had to introduce a fudge gain in the digital path to get the model to line up with the measurement (see below).

Attachment #2 shows the OLGs of the two actuation paths.

  • Aforementioned fudge factor for the digital path is included.
  • For the AO path, I assumed a value of the PDH discriminant at the IMC error point to be 13 kHz/V, per my earlier measurement. 
  • I trawled the elog for the most up-to-date info about the IMC servo (elog9457, elog13696, elog15044) and CM board, to build up the model. 

Attachment #3 and #4 show the model, overlaid with measurements of the loop OLG and crossover TF respectively.

  • No fitting is done yet - the next step would be to add the delay of the CDS system for the digital path, and the analog electronics for the AO path. Though these are likely only small corrections.
  • For the crossover TF - I've divided out the digital filters in the CARM_B filter bank, because the injection is made downstream of it (see Attachment #1).
  • There is reasonably good agreement between model and measurement.
  • I think the biggest source of error is the assumed model for the IMC OLTF.

Attachment #5 shows the evolution of the CARM OLG at a few points in the lock acquisition sequence.

  • "Before handoff" corresponds to the state where the primary control is still done by the ALS leg, but the REFL11 signal has begun to enter the picture via the CARM_B path.
  • "IN2 ramped" corresponds to the state where the AO path gain (=IN2 gain on the IMC servo board) has been ramped up to its final value (+0 dB), but the overall loop gain (=IN1 gain on the CM board) is still low. So this is preparation for high bandwidth control. Typically, the arm powers will have stabilized in this state, but ALS control is still on.
  • "Pre-boost" corresponds to an intermediate state - ALS control is off, but the low frequency boosts have not yet been enabled. I typically first engage some ASC to stabilize things somewhat, and then turn on the boosts.
  • "Final" - self explanatory.

Next steps:

Now the I have a model I believe, I need to think about whether there is any benefit to changing some of these loop shapes. I've already raised the possibility of changing the shape of the boosts on the CM board, with which we could get a bit more suppression in the 100 Hz - 1kHz region (noise budget of laser frequency noise --> DARM required to see if this is necessary). 

  15367   Wed Jun 3 02:08:00 2020 gautamUpdateLSCPower buildup diagnostics

Attachments #1 and Attachments #2 are in the style of elog15356, but with data from a more recent lock. It'd be nice to calibrate the ASDC channel (and in general all channels) into power units, so we have an estimate of how much sideband power we expect, and the rest can be attributed to carrier leakage to ASDC.

On the basis of Attachments #1, the PRG is ~19, and at times, the arm transmission goes even higher. I'd say we are now in the regime where the uncertainty of the losses in the recycling cavity - maybe beamsplitter clipping? is important in using this info to try and constrain the arm cavity losses. I'm also not sure what to make of the asymmetry between TRX and TRY. Allegedly, the Y arm is supposed to be lossier.

Quote:

This is very interesting. Do you have the ASDC vs PRG (~ TRXor TRY) plot? That gives you insight on what is the cause of the low recycling gain.

  15368   Wed Jun 3 02:14:32 2020 gautamUpdateASCPRC ASC improves arm transmission RIN

Summary:

I implemented an ASC servo for the PRC, with the POP QPD as a sensor, and the PRM as the actuator. This has improved the stability of the lock (longer locks are possible), and also reduced the RIN of the arm transmission.

Details:

Attachment #1 shows the in-loop error signal suppression, and some out-of-loop monitors (POP22 and POPDC).

  • To practise and get some workable servo settings, I locked the PRMI with carrier resonant (no ETMs).
  • Then, I compare the beam motion witnessed by the POP QPD with and without the feedback loop enabled.
  • I also look at the spectra of the POPDC and POP22 signals, as out-of-loop proxies, to get an estimate of how much noise is being injected out of band.
  • In this toy study, both the in-loop and out of loop monitors show good performance.
  • However, when repeating the same diagnostics with the PRFPMI locked, I note that while the in-loop suppression looks good, POPDC and POP22 report elevated noise, relative to the PRMI carrier case.
  • I don't have a comparison to the PRFPMI locked with the feedback disabled, because of stability reasons. Plus, for the PRMI, the angular feedforward loops were engaged, but for the PRFPMI traces, they were disabled.
  • Nevertheless, the arm RIN goes down by ~2.5 in RMS, so this is doing something good.

Attachment #2 compares the arm transmission RIN with the PRFPMI locked, with and without PRC ASC. The 3 Hz bump is definitely squished, but I think we can do better yet. 

Attachments #3-5 are in the style of elog15361. No Oplev signals yet, I'll add them soon.

I guess what this means is that the stability of the lock could be improved by turning on some POP QPD based feedback control, I'll give it a shot

  15370   Wed Jun 3 11:20:19 2020 gautamUpdateDetCharSummary pages

Summary:

The 40m summary pages have been revived. I've not had to make any manual interventions in the last 5 days, so things seem somewhat stable, but I'm sure there will need to be multiple tweaks made. The primary use of the pages right now are for vacuum, seismic and PSL diagnostics.

Resources:

Caveats:

  • Intermittent failures of cron jobs
    • The status page relies on the condor_q command executing successfully on the cluster end. I have seen this fail a few times, so the status page may say the pages are dead whereas they're actually still running.
    • Similarly, the rsync of the pages to nodus where they're hosted can sometimes fail.
    • Usually, these problems are fixed on the next iteration of the respective cron jobs, so check back in ~half hour.
  • I haven't really looked into it in detail, but I think our existing C1:IFO-STATE word format is not compatible with what gwsumm wants - I think it expects single bits that are either 0 or 1 to indicate a particular state (e.g. MC locked, POX and POY locked etc). So if we want to take advantage of that infrastructure, we may need to add a few soft EPICS channels that take care of some logic checking (several such bits could also be and-ed together) and then assume either 0 or 1 value. Then we can have the nice duty cycle plots for the IMC (for example).
  • I commented out the obsolete channels (e.g. PEM MIC channels). We can add them back later if we so desire.
  • For some reason, the jobs that download the data are pretty memory-heavy: I have to request for machines on the cluster with >100 GB (yes 💯GB) ! of memory for the jobs to execute to completion. The frame-data certainly isn't so large, so I wonder what's going on here - is GWPy/GWsumm so heavy? The site summary pages run on a dedicated cluster, so probably the code isn't built for efficiency...
  • Weather tab in PEM is still in there but none of those channels mean anything right now.
  • The MEDM screenshot stuff is commented out for now too. This should be redone in a better way with some modern screen grab utilities, I'm sure there are plenty of python based ones.
  • There seems to be a problem with the condor .dag lockfile / rescue file not being cleared correctly sometimes - I am looking into this.
  15371   Wed Jun 3 11:40:56 2020 gautamUpdateLSCLock acquisition update portal

For these initial attempts, I was just trying to transition MICH to REFL55Q. I agree, the PRCL situation may be more complicated.

Which 1f signals are you going to use? PRCL has sign flipping at the carrier critical coupling. So if the IFO is close to that condition, 1f PRCL suffers from the sign flipping or large gain variation.

  15372   Wed Jun 3 18:49:47 2020 gautamUpdateLSCPRG and CARM signal sign

Summary:

I am inclined to believe that the arm cavity losses are such that the IFO is overcoupled. Some calculations, validated with Finesse modeling also suggest that there isn't a sign change for the CARM error signal when the IFO goes from being undercoupled to overcoupled, but I may have made a mistake here?

Details:

  • We’d like to gain some insight into whether the interferometer is undercoupled, critically coupled, or overcoupled. Factors that determine which of these is true include:
    • Arm cavity losses
    • Recycling cavity losses
  • The proxy by which we determine the recycling gain is usually the arm cavity transmission. Assuming T_PRM = 5.637 % according to the wiki, and assuming the arm cavity transmission is normalized to 1 when locked in the POX/POY state, we can say that the PRG is given by G_PRC = TRX × T_PRM, assuming that the (i) the RF sideband fields are perfectly rejected by the arm cavities and (ii) mode-matching efficiency between the input beam and the arm mode is the same as that between the input beam and the CARM mode.
  • Apart from this, the other measurement we have available to us is the buildup of the sideband fields, namely POP22 and POP110. We can compare the values in the PRMI lock vs the PRFPMI to make some inference.
  • I started off with an analytic calculation of the reflectivity of the compound arm cavity mirror.
    • Attachment #1 suggests we will have an over-coupled IFO for arm cavity losses below ~200 ppm, which is a regime we are almost certainly in now.
  • Then, I repeat the analysis for the coupled CARM cavity, with the end mirror as the compound arm mirror and the input mirror as the PRM.
    • I assume 2 % loss in the PRC.
    • Attachment #2 shows that while the carrier field goes through a sign change in amplitude reflectivity (as expected), the sideband fields dont.
    • Per equation 4.2 of Koji's thesis, the error signal for CARM depends on the (signed) IFO reflectivity, and the absolute value of the derivative of the arm cavity reflectivity for the carrier w.r.t. CARM phase.
    • So, we don't expect the REFL11 signal to show a sign change.
    • The situation is more complicated for PRCL in REFL11, because as explicitly evaluated in Eq 4.3 of Koji's thesis, there are two terms that contribute, and their relative magnitudes will dictate the overall sign. 
  • For a Finesse validation, I use a simplified 3 mirror coupled cavity to approximate the PRFPMI. I also retained the RF sidebands for diagnostic purposes. The idea was to study these PRG proxies and what their expected behavior is.
    • Attachment #3 shows the PDH error signal in the (arbitrarily defined) REFL11 I quadrature. While the optical gain changes as a function of the arm cavity loss, the actual slope does not change sign. The fact that the zero crossing doesn't happen at exactly 0 CARM offset is because of higher order mode light at the REFL port (in my model, I tried to preserve the flipped folding mirror situation so the mode matching between the arm cavity and PRC in my model is ~96%).
    • In fact, this may explain why a CARM_B offset is required to do the ALS-->IR handoff - the ALS servo wants to keep the arm offset to zero, but at that point, the PDH error signal isn't zero, and so the two loops end up fighting each other?
    • Attachment #4 is a more detailed study of the recycling gain as a function of arm cavity loss, but now including losses in the recycling cavity.

Conclusions:

  1. I think the arm cavity losses are in the 60-80 ppm round-trip region. I don't see how we can explain the arm cavity transmission of ~350 otherwise.
  2. The fact that REFLDC decreases as the arm transmission increases is because the input beam is getting better matched to the CARM mode, and there is less junk carrier light. 

Thoughts from others?

  15373   Wed Jun 3 19:19:11 2020 gautamUpdateSUSAll watchdogs tripped

This EQ seems to have knocked all suspensions out. ITMX was stuck. It is now released, and the IMC is locked again. It looks like there are some serious aftershocks going on so let's keep an eye on things.

  15376   Thu Jun 4 20:54:40 2020 gautamUpdateSUSMC1 Slow Bias issues

Summary:

I found that there is an issue with the MC1 slow bias voltages. 

Details:

I usually offload the DC part of the output voltage from the WFS servos to the slow bias voltage sliders, so as to preserve maximum actuation range from the fast system. However, today, I found that this servo wasn't working well at all. So I dug a little deeper. Looking at the EPICS database records:

  • The user-facing channels are "PIT" and "YAW" bias voltages.
  • These are converted to voltages to be sent to individual coils by some calc channels in the EPICS database record. So, for example, the voltage to be sent to the "UL" coil (Upper Left, as viewed from the AR side of the optic), is A+B, where A is the "PIT" voltage and B is the "YAW" voltage. Similar combinations of A and B are used for the other 3 face coils.
  • The problem is obvious - if either A or B > 5V, then the requested voltage to be sent to the UL coil is > 10 V, while the Acromag DACs can put out a maximum of 10 V
  • As it happens, with the IFO currently aligned, MC1 is the only optic which faces this problem. 
  • Why has this not been an issue before? In fact, looking at some old data, the "PIT" and "YAW" bias voltages to MC1 were both ~1-2 V in 2018. But I confirmed that something in the region of ~5 V is required from each of the "PIT" and "YAW" channels to bring the MCREFL spot back to the center of the camera, so something has changed the DC alignment of MC1, maybe an earthquake or something? Anyway, with these settings, 2/4 coils are basically saturated, and so we can only move the optic diagonally. 😢 
  • Other coils that have  requested output voltages > 5V (so more than half the range of the DAC) include MC2 LL (5.2V), and ETMX LL and LR (5.5 and 5.8 V respectively).
  • Either a factor of 0.5 should be included in all the EPICS database records, or else, we should make the "PIT" and "YAW" sliders range only from -5 to +5 V, so that this kind of misleading info isn't wasting time.
  15383   Mon Jun 8 18:14:55 2020 gautamUpdateCDSVertex FEs crashed

Summary:

Around 5pm local time, the three vertex FEs crashed. AFAIK, no one was in the lab or working on anything CDS related, so this is worrying.

Details:

  • Reboot script was used to bring all FEs back - only soft reboots were required.
  • The IMC and arms can now be locked.
  • I think combination of burt + SDF would have reverted all the settings as they should be, but if something appears off, it could be that some EPICS value didn't get reset correctly.
  15391   Thu Jun 11 11:48:43 2020 gautamUpdateVACVac failure

There appears to have been some sort of vacuum failure.

ldas-pcdev1 was down, so the summary pages weren't being generated. I have now switched over to ldas-pcdev6. I suspect some forepump failure, will check up later today unless someone else wants to take care of this.

There was no interlock action, and I don't check the vacuum status every half hour, so there was a period of time last night there was high circulating power in the arm cavities when the main volume pressure was higher than nominal. I have now closed the PSL shutter until the issue is resolved.

  15392   Thu Jun 11 16:14:03 2020 gautamUpdateVACVac failure - probable cause is serial comm glitch

Summary:

It looks like the main vacuum interlock was tripped due to a serial communication error from the TP2 controller. With Rana/Koji's permission, I will open V1 and expose the main volume to TP1 again (#2 in last section).

Details:

  • The vacuum interlock log file at /opt/target/vac.log on c1vac suggests that the interlock was tripped because "TP2 is too warm".
  • Looking back at the diagnostics channels, it looks like the TP2 temperature channel registered a rise in temperature of >30 C in <0.2 seconds, see Attachment #1 - seems highly unlikely, probably some kind of glitch in the serial communication? This particular pump is relatively new from Agilent (<2 years installed I think)
  • The PSL shutter was automatically closed at ~1150 am today, see Attachment #2. There is some EPICS logic on c1psl (Acromag server) that checks if C1:Vac-P1a_pressure is greater than 3 mTorr (or greater than 500 Torr for in-air locking of the IMC), in which case it closes the shutter, so this seems consistent with expectations.

Recommended course of action:

  1. Code in some averaging in the interlock code, so that the interlock isn't triggered on some unphysical glitch like this. As shown in Attachment #3, this has been happening for the past 24 hours (though not before, because the interlock wasn't tripped). Probably need the derivative of the temperature as well, and the derivative should be less than 5 C/s or something physical (in addition to the temperature being high) for the interlock to trip.
  2. Re-open V1 to pump down the main volume to nominal pressure so that the interferometer locking activity can resume.
    • One option in the interim is to bypass the TP2 temperature interlock condition.
    • The pressure-based interlocks are probably sufficient to protect the main volume / pumps during the nominal operations - the temperature interlocks are mainly useful during the pumpdown where the TPs have a large load, and so we want to avoid over-stressing them.
  15393   Thu Jun 11 17:35:34 2020 gautamUpdateVACPumpspool UPS needs battery replacement

Summary:

The pumpspool UPS has its "Replace Battery" indicator light on. Might be a good chance to change the UPS, but at the very least, we should put in fresh batteries (last replacement was in Aug 2017).

I'll say this again - the pumpspool area is noisier than I remember it being, I think one / both of the roughing pumps backing TP2 / TP3 need tip-seal replacements.

BTW, EX is 5C hotter than EY, by virtue of the tarnac outside? In fact, judging by Steve's thermometers, EX reports a 12C swing in 24 hours between 30 C and 18 C (so almost no temperature control) while EY reports a 5C swing between 20 and 25 C. This is borne out by the ETM Oplev data I think...

  15397   Fri Jun 12 19:02:52 2020 gautamUpdateVACPumpspool UPS needs battery replacement

I still don't understand why restoring the vacuum is contingent on this functionality working. All the TPs have their own internal logic to shutdown the pump if some damage threshold is exceeded. Plus, we have the pressure-sensor based interlocks to protect the main volume as well as pumps. While the extra redundancy from the readbacks from the controller is useful, clearly it isn't the first line of defense?

The main volume pressure is currently ~10mTorr. If we pump down before this reaches 500mTorr, the procedure is pretty straightforward. Otherwise, we have to do the dance with the manual throttling valve (judging by current rate of increase, unlikely to exceed this over the weekend, but I lose IFO time).

Obviously I don't want to rush this and have some permanent damage, so I'll stay out of this unless otherwise instructed.

  15399   Fri Jun 12 19:33:31 2020 gautamUpdateVACPumpspool UPS needs battery replacement

Didn't mean to sound whiny. I will wait until the vacuum team tells me it is okay.

Quote:

The vacuum safety policy and design are not clear to me, and I don't know what the first and second defense is. Since we had limited time and bandwidth during the remotely-supported recovery work today, we wanted to work step by step.

The pressure rising rate is 20mtorr/day, and turning on TP3 early next week will resume the main-volume pumping without too much hustle. If you need the IFO time now, contact with Jon and use backing with TP3.

  15404   Wed Jun 17 16:27:51 2020 gautamUpdateVACQuestions/comments on vacuum

I missed the vacuum discussion on the call today, but I have some questions/comments:

  • Isn’t it true that we didn’t digitally monitor any of the TP diagnostic channels before 2018 December? I don’t have the full history but certainly there wasn’t any failure of the vacuum system connected to pump current/temp/speed from Sep 2015-Dec2018, whereas we have had 2 interruptions in 6 months because of flaky serial communications.
  • According to the manuals, the turbo-pumps have their own internal logic to shut off the pump when either bearing temperature exceeds 60C or current exceeds 1.5A. I agree its good to have some redundancy, but do we really expect that our outer interlock loops will function if the internal ones fail?
  • In what scenario do we expect that all our pressure gauge readbacks fail, but not the TP readbacks? If so, won’t the differential pressure conditions protect the vacuum envelope, and the TPs internal shutoffs will protect the pumps? Except during the pump down phase perhaps, when we want to give a little more headroom to the small TPs to stress them less?

At the very least, I think we should consider making the interlock code have levels (like interrupts on a micro controller). So if the pressure gauges are communicating and are reporting acceptable pressure readings, we should be able to reject unphysical readbacks from the TP controllers.

I still don’t understand why TP2 can’t back TP1, but we just disable all the software interlock conditions contingent on TP2 readbacks. This pump is far newer than TP3, and unless I’ve misunderstood something major about the vacuum infrastructure, I don’t really see why we should trust this flaky serial readbacks for any actionable interlocks, at least without some AND logic (since temperature, current and speed aren’t really independent variables).

I also think we should finally implement the email alert in the event the vacuum interlock is tripped. I can implement this if no one else volunteers.

This might also be a good reminder to get the documentation in order about the new vacuum system.

  15407   Thu Jun 18 12:00:36 2020 gautamUpdateVACQuestions/comments on vacuum

I agree there were MEDM fields, but I can't find any record of these channels being recorded till 2018 December, so I don't agree that they were being digitally monitored. You can also look back in the elog (e.g. here and here) and see that the display fields are just blank. I would then assume that no interlocks were dependent on these channels, because otherwise the vacuum interlocks would be perpetually tripped.

Looking at images of the old vac screens, the TP2/3 rotation speed and status string were digitally monitored. However I don't know if there were software interlocks predicated on those.

Sorry but I'm having trouble imagining a scenario how the pressure gauges wouldn't register this before the IFO volume is compromised. Is there some back of the envelope calculations I can do to understand this? Since both the pressure gauges and the TP diagnostic channels are being monitored via EPICS, the refresh rate is similar, so I don't see how we can have a pump temperature / speed / current threshold tripped but NOT have this be registered on all the pressure gauges, seems like a bit of a contrived scenario to me. Our thresholds currently seem to be arbitrary numbers anyway, or are they based on some expected backstreaming rate? Isn't this scenario degenerate with a leak elsewhere in the vacuum envelope that would be caught by the differential pressure interlocks?

The temperature and current interlocks are implemented precisely because the pumps can shut themselves off. The concern is not about damaging the pumps (their internal logic protects against that); it's that a pump could automatically shut down and back-vent the IFO to atmosphere. Another interlock (e.g., the pressure differentials) might catch it, but it would depend on the back-vent rate and the scenario has never been tested. The temperature and current interlocks are set to trip just before the pump reaches its internal shut-down threshold.

For the email alert, can you expose a soft channel that is a flag - if this flag is not 1, then the service will send out an email.

That would be awesome if you're willing to volunteer. I agree this would be great to have.
  15410   Thu Jun 18 15:46:34 2020 gautamUpdateVACQuestions/comments on vacuum

So why not just have a special mode for the interlock code during pumpdown and venting, and during normal operation we expect the main volume pressure to be <100uTorr so the interlock trips if this condition is violated? These can just be EPICS buttons on the Vac control MEDM screen. Both of these procedures are not "business as usual", and even if we script them in the future, it's likely to have some operator supervising, so I don't think it's unreasonable to have to switch between these modes. I just think the pressure gauges have demonstrated themselves to be much more reliable than these TP serial readbacks (as you say, they worked once upon a time, but that is already evidence of its flakiness?). The Pirani gauges are not ultra-reliable, they have failed in the past, but at least less frequently than this serial comm glitching. In fact, if these readbacks are so flaky, it's not impossible that they don't signal a TP shutdown? I just think the real power of having these multi-channel diagnostics is lost without some AND logic - a turbopump failure is likely to result in an increase in pump current and temperature increase and pump speed decrease, so it's not the individual channel values that should be determining if an interlock is tripped.

I definitely think that protecting the vacuum envelope is a priority - but I don't think it should be at the expense of commissioning time. But if you think these extra interlocks are essential to the safety of the vacuum system, I withdraw my request.

I don't disagree that the pressure gauges would register the change. What I'm not sure about is whether the change would violate any of the existing interlock conditions, triggering a shutdown. Looking at what we have now, the only non-pump-related conditions I see that might catch it are the diffpres conditions:

It would be better to have a flag channel, might be useful for the summary pages too. I will make it if it is too much trouble.

There's already a channel C1:Vac-error_status, where if the value is anything other than an empty string, there is an interlock tripped. Does that work?
  15415   Fri Jun 19 09:57:35 2020 gautamUpdateVACQuestions/comments on vacuum

For this particular email service, ideally the email should be sent out as soon as the interlock is tripped, so this would require a line of code to be added to the main interlock code. Which I guess would require a restart of the interlock service. So let me know when you guys plan to do the dry-pump tip seal replacement operation (when I presume valves will be closed anyways) so that we can do this in a minimally invasive way.

Quote:

Ok, this can be added pretty easily. Its value will just be toggled between 1 and 0 every time the interlock server raises/clears the existing string channel. Adding the channel will require restarting the whole vac IOC, so I'll do it at a time when Jordan is on hand in case something fails to come back up.

  15418   Fri Jun 19 16:30:09 2020 gautamUpdateASCSome thoughts about ASC

Summary:

In ELOG 15368, I had claimed that the POP QPD based feedback servo actuating on the PRM stabilized the lock. I now believe this scheme of sensing using the POP QPD and feeding back to the PRM is not a good topology for stabilizing the PRC angular motion.

Details:

  • I was never able to get a measurement of the OLTF of this loop that made sense 
    • the loop was initally commissioned with the PRMI locked on the carrier, and the settings hence inferred to give a ~5 Hz UGF loop were used in the PRFPMI lock.
    • In the PRFPMI configuration, however, the loop gain seemed way too low when I measured using the usual IN1/IN2 method.
    • So it is critical for the lock stability that the angular feedforward works well, which it kind of does now (not that I have changed anything, but the glitches in the seismometer have not resurfaced recently).
    • Hopefully, this becomes less of an issue once we replace the TTs with SOS and OSEM based damping.
  • To get some more insight, I did some finesse modeling
    • Attachment #1 shows the sensing response at the QPDs we have available currently (POP and TR). 
    • I included the telescopes (propagation distances, in-air lenses) to these QPDs as best as I could.
    • A simplified model (3 mirror coupled cavity) is used, so there isn't really a common/differential mode in this picture, but we still get some insight I think.
    • Specifically, once the full lock is realized, the PRC optic motion isn't sensed well with our QPDs, and so it was some fluke that turning on these PRC angular feedback loops worked. 
    • Attachment #2 shows the same info as Attachment #1, but with the pendulum transfer functions (and radiation pressure effects) included. The SOS suspensions are modelled as f0=0.7/0.8 Hz (for P/Y), Q=5, while the tip-tilts have f0~5 Hz, Q~10. The high frequency phase is 0 degrees and not 180 as expected because of the pendulum complex pole pair because of the way the quantity is computed in Finesse.
  • The current scheme I use is:
    • DC couple the ITM oplevs, using their individual Oplev QPDs.
    • Use the TR QPDs, mixed to actuate on the ETMs in a common/differential way.
    • I think the system is under-determined with the sensors we currently have - we wan't to sense the 10 angular modes - PIT and YAW for the PRC, Csoft, Chard, Dsoft and Dhard (using the terminology from Kate's thesis), but we only have 6 sensors of the same field (POP, TRX and TRY QPDs, PIT and YAW from each).
    • So we need more sensors?
  • One thing that can easily be improved I think is to make the ASS system work at high power. 
    • think this should be as simple as scaling the gain for the loops to work for the high power.
    • Then we can counteract the input pointing drift at least.
    • But the ITM Oplev DC coupling would need to be turned OFF and then ON again, I'm not sure if this will introduce some transient that will destroy the lock...

I would also like to bring up the topic of implementing some WFS for the interferometer fields again, there doesn't seem to be any mention of this in the procurement/planning for the BHD. It is not obvious to me yet that we need WFS and not just DC QPDs from a noise point of view, but at least we should discuss this.

  15419   Fri Jun 19 17:06:50 2020 gautamUpdateLSCWhat should the short-term commissioning goals be?

Summary:

I want some input about what the short-term (next two weeks) commissioning goals should be.

Details:

Before the vacuum fracas, the locking was pretty robust. With some human servoing of the input beam, I could maintain locks for ~1 hour. My primary goals were:

  1. Transition the vertex length DoF control from 3f signals to 1f signals.
  2. Turn on some MICH-->DARM feedforward cancellation, because the noise between ~100 Hz and ~1kHz is dominated by this cross-coupling.

I didn't succeed in either so far.

  1. I find that there is poor separation of the length DoFs in the 1f sensors, which makes this transition hopeless.
    • Why should this be? I can't get any sensing matrix in Finesse to line up with what I measure in-lock.
    • One hypothesis I came up with (but haven't yet tested) is that the offsets from the 3f photodiodes are changing from time to time, which somehow changes the projection of the various DoFs onto the photodiode quadratures. 
    • The attached GIF shows the variation in the measured sensing matrix on two days - while the sensing of MICH/PRCL in the 3f photodiodes have hardly changed, they are significantly different in the 1f photodiodes. Note that the I and Q have changed for REFL11 and REFL55 between the two days because I changed the demod phase.
    • I also thought that maybe the CARM suppression isn't sufficient for REFL11 to be used as a PRCL sensor - but even after engaging a CM board SuperBoost, I was unable to realize the PRCL 3f-->1f transition, even though the CARM-->REFL11 coupling did get smaller in the measured sensing matrix (red line in the GIF). I don't think we can juice up the CARM gain much more without modifying the CM board boosts, see Attachment #1.
  2. I was able to measure the MICH CTRL --> DARM ERR transfer function with somewhat high coherence (~0.98).
    • I then used the infrastructure available in the LSC model to try and implement some cancellation, but didn't really see any effect.
    • Perhaps the TF needs to be measured with higher coherence.
    • It may also be that if I am able to successfully execute the 3f-->1f transition, the coupling gets smaller because the 1f sensing noise is lower?

I guess apart from this, we want to run the ALS scan to try and infer something about the absorption-induced thermal lens. I guess at this point, the costs outweigh the benefits in trying to bring in the SRC as well, since we will be changing the SRC config?

  15420   Fri Jun 19 19:21:25 2020 gautamUpdateGeneralPSL shutter re-opened

The PSL shutter was closed from the vacuum interlock trip. Today, I did the following:

  • Re-aligned input beam to PMC to recover high transmission / low reflection.
  • Re-set the LSC offsets.
  • ETMX watchdog was tripped. Reset it.
  • Opened the PSL shutter, IMC autolocker was able to lock the cavity almost immediately.
  • Tested POX/POY locking, ran the ASS to maximize single arm transmission.

All looks good for now. I will probably get back to PRFPMI locking Monday.

  15423   Mon Jun 22 17:51:50 2020 gautamUpdateCDSc1iscaux was down

The machine needed a hard reboot as it was un-ssh-able. 

The exact time that the machine went down is unknown because the blinkys were not DQ-ed. I've now added these to the EDCU to make these channels actually useful, and we may look back on the reliability (or otherwise) of the Acromag system. To my memory, this is the ~5th time one of the new Acromag servers has needed a hard reboot. While this may be less frequent (?) than the VME machines, perhaps there is some other reason for these dropouts. Maybe something to do with the martian network?

Anyway the machine is back up and running now.

  15427   Wed Jun 24 17:20:16 2020 gautamUpdateLSCWhat should the short-term commissioning goals be?

Per the discussion at the meeting today, the plan of action is:

  1. Lock the PRMI on carrier and measure the sensing matrix, see if the MICH and PRCL signals look sensible in 1f and 3f photodiodes.
  2. Try locking CARM on POP55 (since there is currently no POP55 photodiode, can we use POX/POY as an intermediary?).
  3. For the ASC, can we hijack one of the IMC WFS heads to study what the AS port WFS signals would look like, and maybe close a feedback loop on the ETMs?
    • My guess is no, because currently, the L2A is so poorly tuned on MC2 that the CARM length control messes with the alignment of the IMC significantly.
    • So we need the IMC WFS loops to maintain the pointing.
    • Of course, the MC2 L2A can be tuned to mitigate this problem. 
    • I also believe there is something funky going on with the WFS heads. More to follow on that in a later elog.
    • Apart from these issues, for this scheme to be tested, some mods to the c1ioo model will have to be made so that we can route the servo output to the ETMs (as opposed to the IMC mirrors as is the usual case).

If I missed something, please add here.

Quote:

Summary:

I want some input about what the short-term (next two weeks) commissioning goals should be.

  15428   Wed Jun 24 22:33:44 2020 gautamUpdateSUSEQ tripped all suspensions

This earthquake tripped all suspensions and ITMX got stuck. The watchdogs were restored and the stuck optic was released. The IFO was re-aligned, POX/POY and PRMI on carrier locking all work okay.

  15431   Thu Jun 25 15:11:00 2020 gautamUpdateSUSMC1 coil driver resistance quartered

I implemented this change today. We only had 100 ohm, 3W resistors in stock (no 200 ohm with adequate power rating). Assuming 10 V is dropped across this resistor, the power dissipation is V^2/R ~ 1 W, so we should have sufficient margin. DCC entry has been updated with new schematic and photo of the component side of the board. Note that the series resistance of the fast actuation path was untouched.

As expected, the requested voltage no longer exceeds the Acromag DAC range, it is now more like 2.5 V. However, I still notice that the MC REFL spot moves somewhat diagonally on the camera image - so maybe the coil gains are seriously imbalanced? Anyway, the WFS control signals can once again be safely offloaded to the slow bias voltages once again, preserving the fast ADC range for other actuation.

The Johnson noise of the series resistor has now increased by a factor of 2, from ~6.4 pA/rtHz to 12.8 pA/rtHz. Assuming a current to force coefficient of 1.6 mN/A per coil, the length noise of the cavity is expected to be 12.8e-12 * 0.064/0.25/(2*pi*100)^2 ~ 8e-18 m/rtHz at 100 Hz. In frequency units, this is 80 uHz/rtHz. I think our IMC noise is at least 10 times higher than this at 100 Hz (in any case, the noise of the coil driver is NOT dominated by the series resistance). Attachment #1 confirms that there isn't any significant MCF noise increase, and I will check with the arm cavity too. Nevertheless, we should, if possible, align the optic better and use as high a series resistance as possible.

The watchdog for MC1 was disabled and the board was pulled out for this work. After it was replaced, the IMC re-locks readily.

Quote:

But this does not solve the MC1 issue. Only we can do right now is to make the output resister half, for example.

  15433   Fri Jun 26 16:53:38 2020 gautamUpdateElectronicsRFPD characterization

Summary:

While the vacuum system was knocked out, I measured the RF transimpedance (using the AM laser setup, didn't do the shot noise intercept current measurement for now) of all the RFPDs (except PMC REFL). At the very least, the following photodiodes are suspect:

  1. WFS heads - expected transimpedance is 50 kohm unattenuated, and 5 kohm attenuated. I measure values that are x10 lower than this, and the segments are significantly imbalanced. Morover, the attenuators for some quadrants appear to do nothing. This could be a problem with the Acromag system I guess, but the measured transimpedance is nowhere close to the "expected" value. See Attachments #1 and #2. You can also see that the response at 55 MHz is significantly attenuated, so I'm guessing trying to measure the AS port ASC sensing response is going to be difficult.

    Note that I assumed a 1kohm DC transimpedance, which is what I expect from the schematic and also is consistent with the DC voltage I measured, knowing the approximate optical power incident on the photodiode.
  2. POP 22/ POP 110 - this is a Thorlabs PDA10CF diode. It should have a flat gain profile out to ~100 MHz, but I measure some weird features. The other PDA10CF we use, at AS110, shows a more reasonable response. See Attachment #3. I don't know what kind of failure mode this is? Anyway I'll try testing another PDA10CF and if it looks more reasonable, I'll switch out this diode. FWIW, the measured AS110 gain is ~3kohms, whereas the datasheet tells us to expect 5 kohms.

For the remaining photodiodes, I measure a transimpedance that is within ~20% of what is on the wiki page. The notches may benefit from some retuning. While I have the data, I will fit this and post a more complete report on the wiki.

Update July 6 1145am: WFS response plots now have legends mapping quadrants, and I've also added the response of a spare PDA10CF (which is now the new POP22/POP110 photodiode).

  15434   Sun Jun 28 15:30:52 2020 gautamUpdateSUSMC1 sat-box de-lidded

Judging by the summary pages, some 18 hours after this change was made and the board re-installed, the MC1 shadow sensors began to report frequent glitches. I can't think of a plausible causal connection, especially given the 18 hour time lag, but also hard to believe there isn't one? As a result, the IMC is no longer able to stay locked for extended periods of time. I did the usual cable squishing, and also took off the lid to see if that helps the situation.

While the reduced series resistance means there is more current flowing through the slow path

  1. There isn't actually an increase in the net current flowing through the satellite box - this change just re-allocates the current from the fast path to the slow path, but by the time it reaches the satellite box, the current is flowing through the same conductor.
  2. afaik, the current buffers on the coil driver aren't overdriven - they are rated for 300 mA. No individual coil is drawing more than 30 mA.
  3. the resistors themselves should be running sufficiently below their rated power of 3W (I estimate 2.5 V ^2 / 100 ohms ~ 60 mW).
  4. The highest current should be through the UL and LR coils according to the voltage outputs from the Acromag. But the UL coil doesn't show significant glitching, and the LL one does despite drawing negligible DC current.

The attached FLIR camera image re-inforces what we already know, that the thermal environment inside the satellite box is horrible. The absolute temperature calibration may be off, but it was difficult to touch the components with a bare finger, so I'd say its definitely > 70 C.

Quote:

I implemented this change today. We only had 100 ohm, 3W resistors in stock (no 200 ohm with adequate power rating). Assuming 10 V is dropped across this resistor, the power dissipation is V^2/R ~ 1 W, so we should have sufficient margin. DCC entry has been updated with new schematic and photo of the component side of the board. Note that the series resistance of the fast actuation path was untouched.

  15436   Sun Jun 28 17:36:35 2020 gautamUpdateSUSMC1 sat-box de-lidded

Hmm I can't seem to export with the colorbar, might be just my phone though. I tried to add some "cursors" with the temperature at a few spots, but the font color contrast is poor so you have to squint really hard to see the temperatures in the photo I attached.

I'll leave the MC1 box open overnight and see if that improves the situation, and if not, I'll switch in the SRM satellite box tomorrow.

Quote:

does the FLIR have an option to export image with a colorbar?

How about just leave the lid open? or more open? I don't know what else can be done in the near term. Maybe swap with the SRM sat box to see if that helps?

  15438   Mon Jun 29 11:55:46 2020 gautamUpdateSUSMC1 sat-box de-lidded

There was no improvement to the situation overnight. So, I did the following today:

  1. Ramped bias voltages for SRM and MC1 to 0, shutdown watchdogs.
  2. Switched SRM and MC1 satellite boxes. The SRM satellite box lid was opened, while the MC1 lid was left open. The boxes have also been re-labelled lest there be some confusion about which box belongs where.
  3. Restored watchdogs and bias voltages. Curiously, the MC1 optic now only requires half the bias voltages it did before to have the correct DC alignment for the optic. The Satellite box is just supposed to be a passive conduit for the drive current, so this is indicative of some PCB traces/cabling being damaged inside what was previously the MC1 satellite box?

IMC is now locked again, I will monitor for glitching/stability.

Update 6pm PDT: as shown in Attachment #1, there is a huge difference in the stability of the lock after the sat box swap. Let's hope it stays this way for a while...

Quote:

I'll leave the MC1 box open overnight and see if that improves the situation, and if not, I'll switch in the SRM satellite box tomorrow.

  15439   Mon Jun 29 15:56:02 2020 gautamUpdateElectronicsRFPD characterization

A more comprehensive report has been uploaded here. I'll zip the data files and add them there too. In summary:

  1. There are several problems with the WFS heads
    • Some attenuators don't seem to work. This could be a problem with the Acromag BIO, or with the relay on the head itself.
    • The measured transimpedance at 29.5 MHz is much lower than expected. We expect ~50 kohms with no attenuation, and 5 kohms without. I measure 100 ohm - 2 kohm with the attenuation disabled, and ~200 ohms with it enabled.
    • Quadrant #3 on both WFS heads behaves differently from the others. There is also evidence of a 200 MHz oscillation for quadrant 3.
    • For some reason, there is a relative minus sign between the TFs measured for the WFS and for the RFPDs. I don't understand where this is coming from - all the OpAmps in the LSC PDs and WFS heads are configured as non-inverting, so why should there be a minus sign? Is this indicative of the polarity of the LEMO output being somehow flipped?
  2. POX 11 photodiode does not have a notch at 22 MHz.
  3. AS55 resonance appears to have shifted closer to 60 MHz, would benefit from a retuning. But the notches appear fine.
  4. PDA10CF photodiode used as the POP22/POP110 readback appears broken in some strange way. As shown in the linked document, a spare PDA10CF in the lab has a much more reasonable response, so I am going to switch out the POP22/POP110 diode with this spare.

I'll upload the data and analysis notebook + liso fit files to the wiki as well shortly. The data, a Jupyter notebook making the plots, and the LISO fit files have been uploaded here.

I didn't do it this time but it'd be nice to also do the noise measurement and get an estimate for the shot-noise intercept current.

Quote:

While I have the data, I will fit this and post a more complete report on the wiki.

  15442   Tue Jun 30 10:59:16 2020 gautamUpdateLSCThree sensing matrices

Summary:

I injected some sensing lines and measured their responses in the various photodiodes, with the interferometer in a few different configurations. The results are summarized in Attachments #1 - #3. Even with the PRMI (no arm cavities) locked on 1f error signals, the MICH and PRCL signals show up in nearly the same quadrature in the REFL port photodiodes, except REFL165. I am now thinking if the output (actuation) matrix has something to do with this - part of the MICH control signal is fed back to the PRM in order to minimize the appearance of the MICH dither in the PRCL error signal, but maybe this matrix element is somehow horribly mistuned?

Details:

Attachment #1:

  • ETMs were misaligned and the PRMI was locked with the carrier resonant in the cavity (i.e. sidebands reflected).
  • The locking scheme was AS55_Q --> MICH and REFL11_I --> PRCL.

Attachment #2:

  • The PRFPMI was locked. The vertex DoFs were still under control using 3f error signals (REFL165_I for PRCL and REFL165_Q for MICH).
  • Still, the MICH/PRCL degeneracy in all photodiodes except REFL165 persists.

Attachment #3:

  • Nearly identical configuration to Attachment #2.
  • The main difference here is that I applied some offsets to the MICH and PRCL error points.
  • The offsets were chosen so that the appearance of a ~300 Hz dither in the length of MICH/PRCL was nulled in the AS110_Q / POP22_I signals respectively.
  • For the latter, the appearance of this peak in the POP110_I signal was also nulled, as it should be if our macroscopic PRC length is set correctly.
  • The offsets that best nulled the peak were 110 cts for PRCL, 25 cts for MICH. The measured sensing response is 1e12 cts/m for PRCL in REFL165_I and 9.2e11 cts/m for MICH in REFL165_Q. So these offsets, in physical units, are: 110 pm for PRCL and 27 pm for MICH. They seem like reasonable numbers to me - the PRC linewidth is ~7.5 nm, so the detuning without any digital offset applied is only 1.5% of the linewidth.
  • Note that I changed the POP22/POP110 demod phases to maximize the signal in the I quadrature. The final numbers were -124 degrees / -10 degrees respectively.
  • Yet another piece of evidence suggesting these were the correct offsets is that the DC value of POX and POY were zero on average after these offsets were applied.
  • However, the MICH/PRCL responses in the 1f REFL port photodiodes remain nearly degenerate.

Some other mysteries that I will investigate further:

  1. While POP22 indicated stable buildup of 11 MHz power in the PRC, I couldn't make any sense of the AS110 signals at the dark port - there was large variation of the signal content in the two quadratures, so unlike the POP signals, I couldn't find a digital demod phase that consistently had all the signal in one of the two quadratures. This is all due to angular fluctuations?
  2. My ASC simulations suggest that the POP QPD is a poor sensor of PRM motion when the PRFPMI is locked. However, I find that turning on a feedback loop with the POP QPD as a sensor and the PRM as the actuator dramatically reduces the low-frequency fluctuations of the arm cavity carrier buildup. 🤔

I blew the long lock last night because I forgot to not clear the ASS offsets when trying to find the right settings for running the ASS system at high power. Will try again tonight...

Quote:

Lock the PRMI on carrier and measure the sensing matrix, see if the MICH and PRCL signals look sensible in 1f and 3f photodiodes.

  15443   Tue Jun 30 22:00:04 2020 gautamUpdateElectronicsGlitchy POX resurfaces

This problem reared its ugly head again. I am inclined to believe the problem is electronic and not on the light, since the POY channels seem immune to this issue (see Attachment #1). I will investigate in the daytime tomorrow. Note that while the POX photodiode head has ~twice the transimpedance than POY (per measurement), the POY signal gets amplified by a ZHL-500-HLN amplifier before heading to the demod electronics (nominal gain is 19dB = x9). There is also some imbalance in the light level at the photodiodes I guess, because overall, the PDH fringe is ~twice as large for the Y arm as the X arm. Basically, the y-axes of the attached plot cannot be directly compared between POX and POY.

Mostly this is an annoyance - right now, the POX signal is only used for locking and dither aligning the X arm cavity, and so once that is done, the locking can proceed (as long as the other channels, e.g. REFL11, aren't glitching as well...)

  15445   Wed Jul 1 12:50:40 2020 gautamUpdatePEMMC1 accelerometers plugged in

I re-connected the 3 accelerometers located near the MC1/MC3 chamber. It was a bit tedious to get the cabling sorted - I estimate the cable is ~80m long, and the excess length had to be wound around a spool (see Attachment #1), which wasn't really a 1 person job. It's neat-ish for now, but I'm not entirely satisfied. I think we should get shorter cables (~20m), and also mount the pre-amp/power units in a rack instead of leaving it on the floor. The pre-amp settings are x100 for all three channels. The MC2 channels are powered, but are unconnected to the seismometers - it was too tedious to unroll the other spool yesterday. Apart from this, the cable for the "Z" channel had to be re-seated in the strain relief clamp.

I did not enable any of the CDS filters that convert the raw signal into physical units, so for now, these channels are just recording raw counts.

Update 7pm: the spectra in the current config are here - not sure what to make of the MC2_Z channel appearing to show lower noise?

Update July 13 2020 430pm: This afternoon, I hooked up the MC2 accelerometer channels too...

  15447   Wed Jul 1 18:16:09 2020 gautamUpdateComputersrossa re-re-revival

In an effort to make a second usable workstation, I did the following (remotely) on rossa today (not necessarily in this order, I wasn't maintaining a live log so I forgot):

  1. Fixed /etc/resolv.conf, so that the other martian machines can be found.
  2. Copied over .bashrc file, and the appropriate lines from /etc/fstab from pianosa to rossa.
  3. Ran sudo apt install nfs-common. Then ran sudo mount -a to get /cvs/cds mounted.
  4. Made symlinks for /users and /opt/rtcds , and /ligo. All of these are used by various environment-setting scripts and I chose to preserve the structure, though why we need so many symlinks, I don't know...
  5. Set up the shell variable $NDSSERVER using export NDSSERVER=fb:8088. I'm not sure how, but I believe DTT, awggui etc use this on startup to get the channel list (any
  6. Followed instructions from Erik von Reis at LHO to install the cds workstation packages and dependencies. Worked like a charm 🎃
  7. As a test, I plotted the accelerometer spectra in DTT, see Attachment #1. I also launched foton from inside awggui, and confirmed that the sample rate is inherited and I could designate a filter. But I haven't yet run the noise injection to test it, I'll do that the next time I'm in the lab.
  8. Also checked that medm, StripTool and ndscope, and anaconda python all seem to work 👍🏾.

So, in summary, rossa is now all set up for use during lock acquisition. However, until this machine has undergone a few months of testing, we should freeze the pianosa config and not mess with it.

Note that this version of the "crtools" is rather new. Please, use them and if there is an issue, report the errors! I am going to occassionally try lock acquisition using rossa. 

Quote:

wiped and install Debian 10 on rossa today

still to be done: config it as CDS workstation

please don't try to "fix" it in the meantime

  15452   Mon Jul 6 00:37:28 2020 gautamUpdateComputersrossa: lib symlink

This is strange - I was definitely able to launch medm when I was working on this machine remotely on Friday. But now, there does seem to be a problem with this shared library being missing.

First of all, I installed mlocate to find where the shared library files are installed. Then I made the symlink, and now sitemap seems to work again.

Weirdly, my changes to /etc/resolv.conf got overwritten somehow. Was this machine rebooted? Uptime suggests it's only been running for ~6 hours at the time of writing of this elog.

sudo apt install mlocate
sudo updatedb
sudo ln -s /usr/lib/x86_64-linux-gnu/libreadline.so.7 /usr/lib/x86_64-linux-gnu/libreadline.so.6
Quote:

when I try 'sitemap' on rossa I get:

medm: error while loading shared libraries: libreadline.so.6: cannot open shared object file: No such file or directory

  15455   Mon Jul 6 12:51:41 2020 gautamUpdateComputersrossa: resolvconf installed

Indeed, this is now fixed by following instructions from here. I rebooted rossa at ~1250 PDT and confirmed that resolv.conf didn't get overwritten. The resolv.conf file also now has the following useful lines at the head:

~>cat /etc/resolv.conf
# Dynamic resolv.conf(5) file for glibc resolver(3) generated by resolvconf(8)
#     DO NOT EDIT THIS FILE BY HAND -- YOUR CHANGES WILL BE OVERWRITTEN
Quote:

yes, I rebooted yesterday to fix the 'steaking white lines' problem in the video/display

maybe we're supposed to edit something besides resolv.conf since that gets over-written on boot for some linux OS

ELOG V3.1.3-