This afternoon I started setting up the Supermicro 5017A-EP that will replace c1vac1/2. Following Johannes's procedure in 13681 I installed Debian 8.11 (jessie). There is a more recent stable release, 9.5, now available since the first acromag machine was assembled, but I stuck to version 8 for consistency. We already know that version to work. The setup is sitting on the left side of the electronics bench for now.
Today I finished setting up the server that will replace the c1vac1/2 machines. I put it on the martian network at the unassigned IP 192.168.113.72. I assigned it the hostname c1vac and added it to the DNS lookup tables on chiara.
I created a new targets directory on the network drive for the new machine: /cvs/cds/caltech/target/c1vac. After setting EPICS environment environment variables according to 13681 and copying over (and modifiying) the files from /cvs/cds/caltech/target/c1auxex as templates, I was able to start a modbusIOC server on the new machine. I was able to read and write (soft) channel values to the EPICS IOC from other machines on the martian network.
I scripted it as a systemd-managed process which automatically starts on boot and restarts after failure, just as it is set up on c1auxex.
Wiring of the power, Ethernet, and indicator lights for the vacuum Acromag chassis is complete. Even though this crate will only use +24V DC, I wired the +/-15V connector and indicator lights as well to conform to the LIGO standard. There was no wiring diagram available, so I had to reverse-engineer the wiring from the partially complete c1susaux crate. Attached is a diagram for future use. The crate is ready to begin software developing on Monday.
All 7 Acromag units are now installed in the vacuum chassis. They are connected to 24V DC power and Ethernet.
I have merged and migrated the two EPICS databases from c1vac1 and c1vac2 onto the new machine, with appropriate modifications to address the Acromags rather than VME crate.
I have tested all the digital output channels with a voltmeter, and some of the inputs. Still more channels to be tested.
I’ll follow up with a wiring diagram for channel assignments.
I've set up a closed subnetwork for interfacing the vacuum hardware (Acromags and serial devices) with the new controls machine (c1vac; 192.168.113.72). The controls machine has two Ethernet interfaces, one which faces outward into the martian network and another which faces the internal subnetwork, 192.168.114.xxx. The second network interface was configured via the following procedure.
1. Add the following lines to /etc/network/interfaces:
iface eth1 inet static
2. Restart the networking services:
$sudo /etc/init.d/networking restart
3. Enable DNS lookup on the martian network by adding the following lines to /etc/resolv.conf:
4. Enable IP forwarding from eth1 to eth0:
$sudo echo 1 > /proc/sys/net/ipv4/ip_forward
5. Configure IP tables to allow outgoing connections, while keeping the LAN invisible from outside the gateway (c1vac):
$sudo iptables -t nat -A POSTROUTING -o eth0 -j MASQUERADE
$sudo iptables -A FORWARD -i eth0 -o eth1 -m state --state RELATED,ESTABLISHED -j ACCEPT
$sudo iptables -A FORWARD -i eth1 -o eth0 -j ACCEPT
6. Finally, because the EPICS 3.14 server binds to all network interfaces, client applications running on c1vac now see two instances of the EPICS server---one at the outward-facing address and one at the LAN address. To resolve this ambiguity, two additional enviroment variables must be set that specify to local clients which server address to use. Add the following lines to /home/controls/.bashrc:
A list of IP addresses so far assigned on the subnetwork follows.
I've completed bench testing of all seven vacuum Acromags installed in a custom rackmount chassis. The system contains five XT1111 modules (sinking digital I/O) used for readbacks of the state of the valves, TP1, CP1, and the RPs. It also contains two XT1121 modules (sourcing digital I/O) used to pass 24V DC control signals to the AC relays actuating the valves and RPs. The list of Acromag channel assignments is attached.
I tested each input channel using a manual flip-switch wired between signal pin and return, verifying the EPICS channel readout to change appropriately when the switch is flipped open vs. closed. I tested each output channel using a voltmeter placed between signal pin and return, toggling the EPICS channel on/off state and verifying the output voltage to change appropriately. These tests confirm the Acromag units all work, and that all the EPICS channels are correctly addressed.
New hardware has been installed in the vacuum controls rack. It is shown in the below post-install photo.
Below is a high-level summary of where things stand, and what remains to be done.
✔ Set up of replacement controls server (c1vac).
✔ Set up of Acromag terminals.
✔ EPICS database migration.
✔ Set up of 16-port IOLAN terminal server (for multiplexing/Ethernetizing the serial devices).
All the serial vacuum signals are now interfaced to the new digital controls system. A set of persistent Python scripts will query each device at regular intervals (up to ~10 Hz) and push the readings to soft channels hosted by the modbus IOC. Similar scripts will push on/off state commands to the serial turbo pumps.
Each serial device is assigned an IP address on the local subnet as follows. Its serial communication parameters as configured in the terminal server are also listed.
The foam in the cable tray wall passage had been falling on the floor in little bite-sized pieces, so I investigated and found a fiber cable that had be chewed/clawed through. I didn't find any droppings anywhere in the 40m, but I decided to bait an un-set trap and see if we'd find activity around it. There has been none so far. If there is still none tomorrow, I will move the trap and keep looking for signs of rodentia. At the moment, the trap is in a box in front of the double doors at the north end of the control room. Next it will be place in the IFO room, up in the cable tray.
gautam: the fiber that was damaged was the one from the LSC rack FiBox to the control room FiBox. So no DAFI action for a bit...
I added lubricating oil to roughing pumps RP1 and RP3 yesterday and this morning. Also, I found a nearly full 5 gallon jug of grade 19 oil in the lab. This should set us up for quite a while. If you need to add oil the the roughing pumps, use the oil in the quart bottle in the flammables cabinet. It is labeled as Leybold HE-175 Vacuum Pump Oil. This bottle is small enough to fill the pumps in close quarters.
The Central Plant building will be undergoing seismic upgrades in the near future. The adjoining north wall along the Y arm will be the first to have this work done, from inside the Central Plant. Project manager Eugene Kim has explained the work to me and also noted our concerns. He assured me that the seismic noise from the construction will be minimized and we will always be contacted when the heaviest construction is to be done.
Tomorrow at 11am, I will bring Mr. Kim and a few others from the construction team to look at the wall from inside the lab. If you have any questions or concerns that you want to have addressed, please email them to me or contact Mr. Kim directly at x4860 or through email at firstname.lastname@example.org .
The schematic of the homodyne configuration is shown below.
Following are the list of components
One set of fiber is now kept along the arm of the interferometer
Fiber coupled (3 No's)
Free space ( 2 No's)
I went through all the elog entries related to CCD calibration. I was wondering if we can use Spectralon diffuse reflectance standards (https://www.labsphere.com/labsphere-products-solutions/materials-coatings-2/targets-standards/diffuse-reflectance-standards/diffuse-reflectance-standards/) instead of a white paper as they would be a better approximation to a Lambertian scatterer.
On calculating the accessible u-v ranges and the % error in magnification (more precisely, %deviation), I got %deviation of order 10 and in some cases of order 100 (attachments 1 to 4), which matches with Pooja's calculations. But I'm not able reproduce Jigyasa's %error calculations where the %error is of order 10^-1. I couldn't find the code that she had used for these calculations and I even mailed her about the same. We can still image with 150-250 mm combination as proposed by Jigyasa, but I don't think it ensures maximum usage of pixel array. Also for this combination the resulting conjugate ratio will be greater than 5. So, use of plano-convex lenses will reduce spherical aberrations. I also explored other focal length combinations such as 250-500 mm and 500-500mm. In these cases, both the lenses will have f-numbers greater than 5. But the conjugate ratios will be less than 5, so biconvex lenses will be a better choice.
Constraints: available lens tube length (max value of d) = 3" ; object distances range (u) = 70 cm to 150 cm ; available cylindrical enclosures (max value of d+v) are 52cm and 20cm long (https://nodus.ligo.caltech.edu:8081/40m/13000).
I calculated the resultant image distance (v) and the required distance between lenses (d), for fixed magnifications (i.e. m = -0.06089 and m = -0.1826 for imaging test masses and beam spot respectively) and different values of 'u'. This way we can ensure that no pixels are wasted. The focal length combinations - 300-300mm (for imaging beam spot), and 100-125mm (for imaging test masses) - were the only combinations that gave all positive values for 'd' and 'v', for given range of 'u' (attachments 5-6). But here 'd' ranges from 0 to 30cm in first case, which exceeds the available lens tube length. Also, in the second case the f-numbers will be less than 5 for 2" lenses and thus may result in spherical aberration.
All this fuss about f-numbers, conjugate ratios, and plano-convex/biconvex lenses is to reduce spherical aberrations. But how much will spherical aberrations affect our readings?
We have two 2" biconvex lenses of 150mm focal length and one 2" biconvex lens of focal length 250mm in stock. I'll start off with these and once I have a metric to quantify spherical aberrations we can further decide upon lenses to improve the telescopic lens system.
Went through all of Pooja's elog posts, her report and am currently cleaning up her code and working on setting up the simulations of spot motion from her work last year. I've also just begun to look at some material sent by Gautam on resonators.
This week, I plan to do the following:
1) Review Gabriele's CNN work for beam spot tracking and get his code running.
2) Since the relation between the angular motion of the optic and beam spot motion can be determined theoretically, I think a neural network is not mandatory for the tracking of beam spot motion. I strongly believe that a more traditional approach such as thresholding, followed by a hough transform ought to do the trick as the contours of the beam spot are circles. I did try a quick and dirty implementation today using opencv and ran into the problem of no detection or detection of spurious circles (the number of which decreased with the increased application of median blur). I will defer a more careful analysis of this until step (1) is done as Gautam has advised.
3) Clean up Pooja's code on beam tracking and obtain the simulated data.
4) Also data like this (https://drive.google.com/file/d/1VbXcPTfC9GH2ttZNWM7Lg0RqD7qiCZuA/view) is incredibly noisy. I will look up some standard techniques for cleaning such data though I'm not sure if the impact of that can be measured until I figure out an algorithm to track the beam spot.
A more interesting question Gautam raised was the validity of using the beam spot motion for detection of angular motion in the presence of other factors such as surface irregularities. Another question is the relevance of using the beam spot motion when the oplevs are already in place. It is not immediately obvious to me how I can ascertain this and I will put more thought into this.
I'm not able to get trends of the TM adjustment test that Rana had asked us to perform, from the dataviewer. It's throwing the following error:
Connecting to NDS Server fb (TCP port 8088)
Server error 7: connect() failed
datasrv: DataWrite failed: daq_send: Resource temporarily unavailable
T0=19-07-20-01-27-39; Length=600 (s)
No data output.
What channels are you trying to read?
Connecting to NDS Server fb (TCP port 8088)
Server error 7: connect() failed
datasrv: DataWrite failed: daq_send: Resource temporarily unavailable
T0=19-07-20-01-27-39; Length=600 (s)
No data output.
Still removing old cable, terminal blocks and hardware. Once new strain reliefs and cable guides are in place, I will need to disconnect cables and reroute them. Please let me know dates and times when that is not going to interrupt your work!
In order to measure the loss in the arm cavities in reflection, we use the DC method described in T1700117.
It was not trivial to find free channels on the LSC rack. The least intrusive way we found was to disconnect the ALS signals DSUB9 (Attachment 1) and connect a DSUB breakout board instead (Attachment 2).
The names of the channels are ALS_BEATY_FINE_I_IN1_DQ for AS reflection and ALS_BEATY_FINE_Q_IN1_DQ for MC transmission. Actually, the script that downloads the data uses these channels exactly...
We misalign the Y arm (both ITM ad ETM) and start a 30 rep measurement of the X arm loss cavity using /scripts/lossmap_scripts/armLoss/measureArmLoss.py and download the data using dlData.py.
We analyze the data. Raw data is shown in attachment 3. There is some drift in the measurement, probably due to drift of the spot on the mirror. We take the data starting from t=400s when the data seems stable (green vertical line). Attachment 5 shows the histogram of the measurement
X Arm cavity RT loss calculated to be 69.4ppm.
We repeat the same procedure for the Y Arm cavity the day after. Raw data is shown in attachment 5, the histogram in attachment 6.
Y Arm cavity RT loss calculated to be 44.8ppm. The previous measurement of Y Arm was ~ 100ppm...
Loss map measurement is in order.
As requested, I placed an order for an assortment of new RF cables: SMA male-male, RG405.
They're expected to arrive mid next week.
I re-ordered the below cables, this time going with flexible, double-shielded RG316-DS. Jordan will pick up and return the RG-405 cables after the holidays.
Our favorite (flexible) RF cable is Belden's 1671J (Jacketed solder-soaked coax cable). It is compatible RG405. I'm not sure if there is off-the-shelf SMA cables with 1671J.
After the disk system trouble, we could not make the RTS running at the nominal state. A part of the troubleshoot FB1 was rebooted. But the we found that the GPS time was a year off from the current time
controls@fb1:/diskless/root/etc 0$ cat /proc/gps
controls@fb1:/diskless/root/etc 0$ date
Thu Sep 2 18:43:02 PDT 2021
controls@fb1:/diskless/root/etc 0$ timedatectl
Local time: Thu 2021-09-02 18:43:08 PDT
Universal time: Fri 2021-09-03 01:43:08 UTC
RTC time: Fri 2021-09-03 01:43:08
Time zone: America/Los_Angeles (PDT, -0700)
NTP enabled: no
NTP synchronized: yes
RTC in local TZ: no
DST active: yes
Last DST change: DST began at
Sun 2021-03-14 01:59:59 PST
Sun 2021-03-14 03:00:00 PDT
Next DST change: DST ends (the clock jumps one hour backwards) at
Sun 2021-11-07 01:59:59 PDT
Sun 2021-11-07 01:00:00 PST
Paco went through the process described in Jamie's elog [40m ELOG 16299] (except for the installation part) and it actually made the GPS time even strange
controls@fb1:~ 0$ cat /proc/gps
I decided to remove the gpstime module and then load it again. This made the gps time back to normal again.
controls@fb1:~ 0$ sudo modprobe -r gpstime
controls@fb1:~ 0$ cat /proc/gps
cat: /proc/gps: No such file or directory
controls@fb1:~ 1$ sudo modprobe gpstime
controls@fb1:~ 0$ cat /proc/gps
I gave some of the data analysts a look around because they asked and nothing was currently going on in the 40m. Nothing was changed.
I have designed new cable supports for the new ribbon cables running up the side of the tables in the vacuum chambers. This is a ribbon cable version of the existing cable clamp cf D010120
The clamps that I have designed (shown in basic sketch attachment 1) will secure the cable at each of the currently used cable supports.
The support consists of a backplate and a frontplate. The backplate is secured to the leg of the table using a threaded screw. The frontplate clamps the cable to the backplate using two screws: one on either side. Between two fascinating points, the cable should have some slack. This should keep the cable from being stiff and help reduce the transfer of seismic noise to the table.
It is possible to stack multiple cables in one of these fasteners. Either you can put two cables together and clamp them down with one faceplate or you can stack multiple faceplates with one cable between each faceplate. in this case the stack would go backplate then cable then faceplate then cable then the second faceplate. this configuration would require longer screws.
The exact specifics about which size screws and which size plates to use still have not been measured by me. But it will happen
We came this morning and noticed the FSS_FAST channel was moving very rapidly. Short inspection showed that MC3 watchdog got tripped. After reenabling the watchdog the issue was fixed and the MC is stable again.
There is a spike in the seismometers 8 hours ago and it was probably due to the 4.2 magnitude earthquake that happened in Malibu beach around that time.
The fan behind the PSL controller is injecting excess band limited noise
While doing noise hunting to improve the BHD lock stability, we noticed peculiar noise bumps in the BH44 error point near (but not exactly at) the even line-harmonics (for example 120 Hz, 240 Hz, ...). Other channels such as C1:HPC-BHDC_SUM, C1:LSC-PRCL_IN1, or even C1:LSC-XARM_IN1 didn't show these features, so we looked at C1:IOO-MC_F_DQ (which represents the free running laser noise above 100 Hz) and to our surprise found this excess noise!
Noise hunting around PSL
Since this noise is present upstream, we decided to hunt around using C1:IOO-MC_F_DQ. We set up a diaggui measurement to do some "live demodulation" as suggested by Koji in order to understand the nature of this noise. In order to get some "video bandwidth" we set up a power spectrum measurement from 114 Hz to 140 Hz (to monitor the usual 120 Hz line noise peak) with a bandwidth of 1 Hz. A single exponential average gave us the 1 Hz narrow spectrum in "real time", from which we noticed its nonstationary character. The band limited excess noise is the result of a peak hovering in the range of 125 to 131 Hz. With this diagnostic set up, we started hunting for its source.
To show the impact in the complete noise spectrum, we took a 10 fixed average measurement with and without the fan being on. The result is in Attachement #2. The spectra are shown along with their rms, which is significantly reduced when the fan is off near the 100 Hz frequency band (where these bumps appear). Anyways, we have left the fan on because the PSL controller needs it so the problem remains, but we have at least identified the source.
Restarted the NDS2 script on Megatron following the instructions here
1) SSH to megatron - "ssh megatron"
2) Switch to nds2mgr using "sudo su nds2mgr"
3) Stop and restart the service using
"sudo /etc/init.d/nds2 stop
sudo /etc/init.d/nds2 start"
After lots of burtrestores and alignment by Koji. The interferometer is live again.
The X/Y green beams are resonating. The X/Y arms were locked with IR. We saw the MI fringe on the POP. However, the AS spot is not visible. Needs to check the optical table?
The auto-restoration process did not bring back all the settings (incomplete restoration), and we had to restore them manually. We took some time to realize it.
Software / Control
On top of these, the Acromag unit should be brought to the workbench, and the upgrade needs to be done. [Moved to the workbench needs elog]
ETMX Eurocard Crate investigation ~ Summary
- The eurocard crate was not properly grounded, which had created a weird powering condition.
- Some busted components on the QPD whitening were replaced. The channels were tested to be OK.
- The OPLEV IF board is functioning as expected
- BUT the OPLEV head seemed busted.
I came in on Sun evening to work on the Eurocard crate.
- Checked the power supply situation with an extender board while no other circuits were mounted on the crate. The +/-20V and +/-15V were supplied, and the corresponding LEDs were lit.
- Then I inserted QPD whitening board (D990399), which made the +15V LED dim... and something on the QPD whitening board burnt off! It was a tantalum cap on the daughterboard. (The daughter board was for the end transmon LPF)
- I checked OPLEV I/F D010033 and this also made +15V LED dim. Something was clearly wrong
- The voltages on the extender board were measured: it turned out that the ground level of the modules (measured at the connector shells) was about 3V off from 0V, which is defined by the Sorensens gnd terminal (which is connected to the rack).
Immediately turned off the +/-15V system and measured the ground resistance between these two grounds. It was ~350Ohm. Does this mean there was an 8~9mA current through somewhere? The R between the Sorensens and the rack frame was basically 0 Ohm.
So it looked like the Eurocard crate was not properly grounded.
- I traced the side panels and found that the ground coming from Sorensen was disconnected, and the ground wire for the crate was isolated. I borrowed a ground slot of the vacant fuse slots (Attachment 1). This made the R between a connector shell to the rack properly ~0Ohm and the powering situation looked finally OK to me.
We now need to check the boards before putting them back in. The QPD Whitening Board (D990399-B) was inspected on the workbench.
- First of all, the daughterboard was isolated.
- The +/-15V supply was connected to the board via an extender board. The board was energized correctly.
- A 1kHz 60mVpp signal was injected from the input nodes. This should make the output voltages to be ~11Vpp.
- It seemed that CH3 had a large (~2V) voltage, and the final output was 0V. Two AD620s for U4 and U21 were replaced. This made the CH3 work identically to the other 3 channels.
- Finally, the daughterboard was taken care. A Ta cap was soldered. The daughterboard was engaged to the power supply nodes, and there seemed to be no problem.20mVpp test input resulted in 320mVpp output.
As Ta Caps are very weak to the reverse and overvoltages, This indicated that there seemed to be some reverse voltage applied to the system.
The board was inserted to the Eurocard crate and now we see a reasonable end QPD transmon signals. (The SUM gain matched with the end transmon PD signal).
Next, I stepped into the OPLEV I/F board. The test indicated that the board was okay.
Then the board was inserted into the crate. It made no problem. And then, the QPD cable was connected => The +15V was drained. So it seemed that the QPD HEAD has an issue. I'll check the head later this week.
Andrei asked me yesterday for the location of the ambient temperature sensor for the X-end. He mentioned that he needed to run so tests because there were ~80° spike over the weekend. After I walked him over, I noticed that the sensor itself was crooked. We may have hit it when we were removing all the ancient cabling from the rack. I removed the sensor and soldered the sensor back on. The Temperature monitor is now working without an issue.
ETMX OPLEV QPD was fixed.
- What we knew was that the QPD head drains +15V when it was connected to the I/F board.
- Suspected the reverse voltage / blown a tantalum cap
- The head position was marked by bases (Attachment 1).
- The head unit was opened. The PCB is D980325 (Attachment 2) and transimpedance R seemed 97.6k. The OPamp is OP497 (quad)
- Removed both tantalum caps. Checked the insulation. One still has a good insulation and the other has short circuit. Yes, this is a typical failure mode of Ta caps.
- Replaced them with two 100uF electrolytic caps (50V rating) (Attachment 3)
- We tested the head unit with the phone flashlight on/off for all four segments. They responded well. So we declared the head was functioning (on the workbench).
- The head was brought back to the Xend table. Positioned it with the bases.
- The ETMX Oplev screen indicated that the head was functioning well with the ambient light level. (i.e. low but about equal light level on the segments)
At the time, there was no oplev light coming into the QPD. So we shined it with phone lights to see the response.
All the segments were responding, so we declared that the oplev signals are nicely transmitted in the CDS.
ETMX Sat amp (D080276) CH3 PD receiver circuit fixed.
At today's meeting, Yuta reported that ETMX LL OSEM had no response.
Mayank and I opened the sat amp for the investigation. Mayank's observation was that the LEDs on the video screen were lit, indicating the LED drivers are all good. Also, he found that the CH3 output (corresponds to LL PD) on the PDmon is +15V saturated.
We checked the I-V amp output (TP6) and it was zero as expected. However, the whitening out (TP8) was saturated at +15V. Once the chip U3 was removed, the monitor and the DAQ out went back to zero.
The chip U3 was replaced with a spare AD822, but this didn't solve the issue. So something was wrong with the whitening stage.
We carefully inspected the board and suspected R24 was not well soldered. This R24 had been originally 19.6k and was replaced by 4.99k at the 40m to bring the whitening pole/zero frequencies to the 40m standard. (See https://dcc.ligo.org/LIGO-S2100744 as an example)
The R24 was soldered again and we replaced AD822 too. This solved the saturation issue.
Mayank returned the unit on the rack and reported that the LL OSEM was responding.
I wanted to check the oplev functionality but the IFO was not well aligned. So started the CDS recovery.
Here is the procedure
- Played a CDS restarting whack-a-mole: a host suffered from DACKILL (DK), so reboot the machine. And this made the other machine fell into DK.
- I decided to power cycle affected machines: sudo-shutdowned c1lsc/c1sus/c1sus2/c1iscex
- Went to the racks and power cycled their IO chassis.
- Powered up these hosts.
After a while, all the RTS came up with no DK error (very good)
Then, burt fest. I wasn't sure what to be restored, but I tried to burtrestore as many snapshots as possible. The snapshots at 20:19 on Jun 21, 2023 were used.
- IMC / WFS is now working well.
- Someone restored the AS/BHD beam spot in the daytime (thanks)
- Both arms are locked and aligned by ASS.
- The green beams are flashing in the arms
- As a consequence of omitting the PENTEK amp (?), the oplev signals (C1:SUS-ETMX_OL_SEGx_IN1, etc.) are now negative. Gave -1 to C1:SUS-ETMX_OL_SEGx_GAIN to compensate it.
- The ETMX damping seemed weak, so C1:SUS-ETMX_SUSPIT_GAIN and C1:SUS-ETMX_SUSYAW_GAIN were doubled to 4.0 and 7.0.
- The oplev calibrations are probably largely off. The servo gains were reduced to 0.01 to have a reasonable servo response. (OLTF to be inspected)
- I don't know how long it was like that, but the servo inputs were off. And the oplev servo gains of -5 was way too big. They were reduced to -1.
ETMX Transmon PD/QPD setting
- The high gain PD gains were recalibrated to be ~1. All the gains were put on C1:SUS-ETMX_TRX_GAIN, and C1:LSC-TRX_GAIN was returned to 1. (previously 15)
- The transmon QPD gain (C1:SUS-ETMX_QPD_SUM_GAIN) was trimmed to be 0.075 to make it equivalent to the high gain PD output.
[Hiroki, Mayank, Koji]
The X end PZT driver for green alignment:
Yesterday evening, Paco and I aimed to:
Once e-FPMI was locked (POX + POY --> CARM_A), we fed the ALS beatnote error signals to CARM_B and slowly mixed CARM_A and CARM_B. ALS control of CARM was successful.
The final values used in C1LSC_AUX_ERR_MTRX were (-0.3 ALSX + 0.3 ALSY) --> CARM_B. Note that these signs depend on the sign of each beatnote. The sign of ALSY could be determined by giving an offset, but without an Acromag we had to use trial and error for the sign of ALSX. We observed that using 0.5 magnitude for each signal resulted in too high of a CARM UGF, making the loop unstable. The magnitudes were reduced to 0.3 to give us a comparable UGF to POX/POY control of CARM.
(-0.3 ALSX + 0.3 ALSY) --> CARM_B
The final ALS CARM OLTF can be found in Attachment 1. Some "wobblyness" was observed in the OLTF. Attachment 2 shows the suppressed in-loop CARM_B error and the out-of-loop CARM_A error. We couldn't identify why CARM_A error has a notch ~325 Hz; this is also present when closing the loop with CARM_A.
We tried to add an LSC CARM offset (would push the PSL frequency away) but could not see the transmission in the arms drop.
Increase stability of ALS CARM, turning loop gains
Achieve a CARM offset maintaining lock
Then proceed to lock PRMI sidebands and reduce the CARM offset for PRFPMI
Tonight I managed to lock CARM and DARM under ALS control only
ALS error signal tuning
To find the error signals for CARM/DARM, I turned on the oscillators (at 307.8 and 313.31 Hz respectively) with 150 counts and enabled FM10 (Notch for sensing matrix) in the CARM and DARM servo banks. I then removed the ALS offsets (C1:LSC-ALSX_OFFSET, C1:LSC-ALSY_OFFSET) and looked at the transfer functions shown in Attachment #2. I optimized the ALS blending until I maximized the CARM and DARM A to B paths and minimized CARM and DARM cross couplings. The signs were chosen to leave a phase of 0.
After measuring the OLTFs for eCARM and eDARM (loop closed with the A error point) and tuning the ALS error signals, I gradually blended the A and B paths and checked the OLTFs for CARM and DARM. During this I realized I needed to disable some of the notch violin filters because they sometimes made the DARM loop unstable after >50% blending. In the end the simultaneous CARM_A/DARM_A to CARM_B/DARM_B handoff was successful in 0.5 seconds. Attachment #3 shows the OLTFs under ALS control.
After getting nominally stable ALS control, I tried adding an offset. The LSC CARM offset range was insufficient, so I ended up directly scanning the C1:LSC-ALSX_OFFSET and C1:LSC-ALSY_OFFSET. The first couple of attempts the ramp time was set to 2.0 seconds, and a step of 0.01 was enough to break the lock. I managed to hold the control with as much as C1:LSC_CARM_A_IN1 offset by ~ 500 (rms ~ 200 counts). I roughly estimate this to be ~ 5% of the CARM pole which is 4 kHz in this case so overall 200 Hz which is not that large.
[JC, Paco, Radhika, Murtaza]
1. WFS Relief
We tried to change the gain to offload the offsets in the reliefMCWFS script but it The gains might need some tuning to get it to work
2. WFS Error Signals Diverging
The error signal C1:IOO-WFS2_I_PIT_MON was staying at a constant offset from 0 using the existing output matrix (C1IOO_WFS_OUTMATRIX). We tried changing the matrix coefficients that may have caused this behavior but it led to divergence in other signals (C1:IOO-WFS1_I_PIT_MON, C1:IOO-WFS2_I_YAW_MON). THIS NEEDS TO BE FIXED
3. BS was aligned using OPLEV readouts and the damping filters were checked. No funny business for BS anymore.
4. The original IFO Align scripts used the suffix "COMM" for each optic. This was changed to "OFFSET" for all arguments by editing the IFO_ALIGN screen (left click-> Execute -> Edit this screen -> !Align -> Label/Cmd/Args)
5. The OPLEV gains for ITMY were unstable and needed some tuning. New gains: C1:SUS-ITMY_OL_PIT_GAIN (14->3.5) and C1:SUS-ITMY_OL_YAW_GAIN (-8->-4). (The upgrade should not have affected this so this could be revisited later).
6. YARM (transmission ~ 1) and XARM (transmission ~ 0.6) were locked successfully!
ITS A GOOD IDEA TO HAVE PRM AND SRM MISALIGNED WHILE TRYING TO LOCK THE IFO.
[Paco, JC, Murtaza]
To fix the WFS loops, went through the following steps
With the WFS loop turned off
- We manually aligned the optics MC1, MC2 and MC3 (IOO -> C1IOO_MC_Align) to maximize transmission (MC Trans Sum -> ~13300)
- We manually aligned the QPDS for WFS1 and WFS2 to center the beam by looking at the DC signals (C1IOO_WFS_QPD). The laser was clipping on WFS2
With the WFS loop turned on
- We changed the gains of the WFS filters for all signals (1.0 -> 2.0), this led to faster conversion but clipping on C1:IOO-WFS2_YAW_OUTPUT. The gains were restored to 1.0 and thus left unchanged.
- We increased the reliefMCWFS gain by a factor of 10 by changing the arguments (Execute -> Edit this screen -> Actions -> label/cmd/args -> Arguments) (0.02 -> 0.2)
- We ran WFSoutMatBalancing.py (17334) to calculate the new output matrix
While aligning today I realized the cavAlign step sizes and step factors had not been updated after the upgrade.
Here are the new MEDM command arguments to launch cavAlign. Only factors not equal to 1 are listed. The updates made by me are in red
[Rana, Radhika, Murtaza]
- We turned on the WFS loops with very small gain (0.01) to see how the error signals behave. There is an existing template to look at the error signals in ndscope (users->Templates->ndscope->IOO->WFS->WFS-overview.yml). We observed C1:IOO-WFS2_IY_DQ stay at a constant offset as we increased the gain to 1.
- The output matrix for WFS (C1IOO_WFS_INMATRIX) was restored to the original value using burtgooey to mitigate the WFSoutMatBalancing.py change 17874.
- (TODO) WFS1 and WFS2 are slightly misaligned as seen on the C1IOO_LOCKMC screen. These need to be aligned when the IFO is unlocked so that the beam is centered on them.
- (TODO) With the PSL shutter turned off, WFS heads should show 0 reading which is not the case. This needs to be corrected for to mitigate the offset readings.
- The electronics upgrade should ideally only affect the suspensions (everything upstream should not need any changes).
- Note: The MC_TRANS error signals look very small in PIT and YAW.
This is a simple representation of the schematic:
gnd# |# Cw2# |# n23# |# Lw2# |# n22# |# Rw2 # | |\ # n2- - - C2 - n3 - - - - | \ # | | | | |4106>-- n5 - Rs -- no# iinput Rd L1 L2 R24 n6- | / | |# nin - | | | | | |/ | Rload # Cd n7 R22 gnd | | | # | | | | - - - R8 - - gnd # gnd R1 gnd R7 # | |# gnd gnd# ##
I chose the values of the components in a realistic way, that is using part available from Coilcraft or Digikey.
Using LISO I simulated the Tranfer Function and the noise of the circuit.
I'm attaching the results.
I'll post the 55MHz rfpd later.
oops, forgotten the third attachment...
here it is
# Resonant RF diode front end
I read a few datasheets of the C30642GH photodiode that we're going to use for the 11 and 55 MHz. Considering the values listed for the resistance and the capacitance in what they define "typical conditions" (that is, specific values of bias voltage and DC photocurrent) I fixed Rd=25Ohms and Cd=175pF.
Then I picked the tunable components in the circuit so that we could adjust for the variability of those parameters.
Finally with LISO I simulated transfer functions and noise curves for both the 11 and the 55MHz photodiodes.
I'm attaching the results and the LISO source files.
Use 10 Ohms for the resistance - I have never seen a diode with 25 Ohms.
p.s. PDFs can be joined together using the joinPDF command or a few command line options of 'gs'.
I spent some time trying to understand how touching the metal cage inside or bending the PCB board affected the photodiode response. It turned out that there was some weak soldering of one of the inductors.
Hartmut suggested a possible explanation for the way the electronics transfer function starts picking up at ~50MHz. He said that the 10KOhm resistance in series with the Test Input connector of the box might have some parasitic capacitance that at high frequency lowers the input impedance.
Although Hartmut also admitted that considering the high frequency at which the effect is observed, anything can be happening with the electronics inside of the box.
I upgraded the old REFL199 to the new REFL55.
To do that I had to replace the old photodiode inside, switching to a 2mm one.
Electronics and optical transfer functions, non normalized are shown in the attached plot.
The details about the modifications are contained in this dedicated wiki page (Upgrade_09 / RF System / Upgraded RF Photodiodes)
These are the dark noise spectrum that I measured on the 11MHz and 55MHz PD prototypes I modified.
The plots take into account the 50Ohm input impedance of the spectrum analyzer (that is, the nosie is divided by 2).
With an estimated transimpedance of about 300Ohm, I would expect to have 2-3nV/rtHz at all frequencies except for the resonant frequencies of each PD. At those resonances I would expect to have ~15nV/rtHz (cfr elog entry 2760).
I have to figure out what are the sources of such noises.
After adding an inductor L=100uH and a resistor R=10Ohm in parallel after the OP547A opamp that provide the bias for the photodiode of REFL11, the noise at low frequency that I had observed, was significantly reduced.
See this plot:
A closer inspection of the should at 11MHz in the noise spectrum, showed some harmonics on it, spaced with about 200KHz. Closing the RF cage and the box lid made them disappear. See next plot:
The full noise spectrum looks like this:
A big bump is present at ~275MHz. it could important if it also shows up on the shot noise spectrum.