Added the conlog directory to the SVN, minus the enormous data directory. We are now free to make changes to the conlog code.
I restarted the conlogger on op340m. This needs to be done when op340m is rebooted--it wasn't done for some reason and so we've lost several days of controls records.
I shorted the inputs on three channels and the outputs on three channels of the Guralp box, and I did similar things with the accelerometers. I was going to move the instruments themselves back, but I didn't have time, so they are still in the box in the corner. If the setup could stay as-is for at least a few hours, that would be awesome.
It was fine when I came in earlier today, but I just got back from dinner, and it's not good. I looked in dataviewer, and it seems to have been sliding out for the past couple of hours... Here is a picture:
I swear I am not responsible this time... all I've been doing is working in the control room.
Mode cleaner bounced back on its own about 2 hours ago.
As Rob noted last Friday, the UPS which powers the Vacuum rack failed. When we were trying to move the plugs around to debug it, it made a sizzling sound and a pop. Bad smells came out of it.
Ben came over this week and measured the quiescent power consumption. The low power draw level was 11.9 A and during the reboot its 12.2 A. He measured this by ??? (Rob inserts method here).
So what we want is a 120 V * 12.2 A ~ 1.4 kVA UPS with ~30-50% margin. We look for this on the APC-UPS site:
On Monday, we will order the SUA2200 from APC. It should last for ~25 minutes during an outage. Its $1300. The next step down is $200 cheaper and gives 10 minutes less uptime.
I was told that, as of last weekend, we now have the capability to save full data for a month, whereas before it was something like 3 days. However, my attempts to get the data from the accidentally-shorted EW2 channel in the Guralp box have all been epic failures. My other data is okay, despite my not saving it for several days after it was recorded. So, my question is, how long can the data actually be saved, and when did the saving capability change?
The EUCLID-style Michelson readout is on the SP table now and is aligned. See image below. I took several power spectra with the plotter attached to the HP3563 (not sure if there's another way to get the data out) and I'm still waiting to calibrate (since dP/dL isn't constant as it isn't locked, this is taking a bit longer). When put into XY mode on the oscilliscope (plotting Voltage at PD2 on the x and Voltage at PD3 on the y), a Lissajous figure as in the first plot below. It's offset and elliptical due to imperfections (noise, dc offset, etc) but can ideally be used to calculate the L_ target mirror movement. By rotating the first quarter wave plate by ~80.5deg counter-clockwise (fast axis was originally at Pi/8, now at 103deg), I was able to turn the Lissajous figure from an ellipse into a more circular shape, which would ideally allow for us to use a circular approximation (much simpler) in our displacement calculations.
Rather than make a new elog post every time I move something, I'm going to just keep updating this Google spreadsheet, which ought to republish every time I change it. It's already got everything I've done for the past week-ish. The spreadsheet can be accessed here, as a website, or here, as a pdf. I will still post something nightly so that you don't have to search for this post, but I wanted to be able to provide more-or-less real-time information on where things are without carpet-bombing the elog.
When I came in earlier today, I noticed that c1susvme2 was red on the DAQ screens. Since the vme computers always seem to be happier as a set, I hit the physical reset buttons on sosvme, susvme1 and susvme2. I then did the telnet or ssh in as appropriate for each computer in turn. sosvme and susvme1 came back just fine. However, I couldn't cd to /cvs/cds/caltech/target/c1susvme2 while ssh-ed in to susvme2. I could cd to /cvs/cds, and then did an ls, and it came back totally blank. There was nothing at all in the folder.
Yoichi showed me how to do 'df' to figure out what filesystems are mounted, and it looked as though the filesystem was mounted. But then Yoichi tried to unmount the filesystem, and it claimed that it wasn't mounted at all. We then remounted the filesystem, and things were good again. I was able to continue the regular restart procedure, and the computer is back up again.
Recap: c1susvme2 mysteriously got unmounted from /cvs/cds! But it's back, and the computers are all good again.
nodus was rebooted by Alex at Fri Aug 14 13:53. I launched elogd.
./elogd -p 8080 -c /export/elog/elog-2.7.5/elogd.cfg -D
It looks like Alex also rebooted all of the control room computers. Or something. The alarm handler and strip tool aren't running.....after I fix susvme2 (which was down when I got in earlier today), I'll figure out how to restart those.
Alex switched the mount point for /cvs/cds on Linux1 to the 1.5 TB RAID array after he finished copying the data from old drives. This required a reboot of linux1, with all the resulting /cvs/cds mount points on the other computers becoming stale. Easiest way to fix that he found was to do a reboot of all the control room machines. In addition, a reboot fest should probably happen in the near futuer for all the front end machines since they will also have stale mount points as well from linux1.
The 1.5 TB RAID array mount is now mounted on /home of linux1, which was the old mount point of the ~300 GB drive. The old drive is now at /oldhome on linux1.
./elogd -p 8080 -c /export/elog/elog-2.7.5/elogd.cfg -D
./elogd -p 8080 -c /export/elog/elog-2.7.5/elogd.cfg -D
The RAID array servicing the Frame builder was finally switched over to JetStor Sata 16 Bay raid array. Each bay contains a 1 TB drive. The raid is configured such that 13 TB is available, and the rest is used for fault protection.
The old Fibrenetix FX-606-U4, a 5 bay raid array which only had 1.5 TB space, has been moved over to linux1 and will be used to store /cvs/cds/.
This upgrade provides an increase in look up times from 3-4 days for all channels out to about 30 days. Final copying of old data occured on August 5th, 2009, and was switched over on that date.
I put all three seismometers and all six accelerometers together in the foam box with peanuts. Three of the accelerometers are facing in the x-direction and three are in the y-direction. Both Guralps are aligned on the NS axis and the Ranger is pointing vertically.
**EDIT: The accelerometers are in the x and z directions, not x and y. Sorry, I was sleepy when I wrote this.**
One of the accelerometers was refusing to show anything, and after a few hours of checking connections and swapping cables, I discovered that someone had unplugged the cable from the ADC. A quick glance in the dataviewer shows that the channel has been unplugged since about 3 in the afternoon on August 8th (Saturday). So... obviously all the accelerometer measurements made with that channel since then did not actually get recorded. Yay.
Anyway, as of 2:45, everything is working and taking data. Clearly we're not getting a full night's worth... hopefully that's okay.
There are two new Matlab files on the svn in /mDV/extra/C1. 'mycsd.m' is to calculate the cross-spectral density between two channels, 'csd_40T_40T_SS1.m' calls this function with the available seismic channels and derives a self-noise spectrum for the vertical axis using all three seismometers. The method requires that there are no correlations between two instruments only which is a bad idealization for certain frequencies if you have seismometers of totally different types.
'mycsd.m' uses the high-gain, low-resolution Nuttall window (built-in Matlab function 'nuttallwin.m'). High-gain windows are used for broad-band spectra like seismic spectra, but it should be exchanged by another window if you plan to look at small-bandwidth features like peaks.
Since the three-channel analysis does not require knowledge of response functions, it could be used to evaluate the performance of the adaptive filter. For example, if three channels responding to the same signal are available, then the ratio of any two csds corresponds to one of the relative transfer functions. You can then compare this function with the result produced by the adaptive filter.
Rana, Jan, Jenne
We noticed that the Ranger data was all bogus at low frequencies. So we checked it and found that the proper procedure had not been used when changing it from horizontal to vertical last week. So the huddle test data from the weekend is not valid for the ranger; we will have to repeat it sometime.
So we used the manual, and extended the hanger rod on top of the Ranger to free the mass. It now has good response and coherence with the Guralps down to 0.1 Hz. See attached plot soon.
When Rob and I were getting started on locking for the evening, Mode Cleaner lost lock a few times, but every time it lost lock, it took forever to reaquire, and was pretty insistent on locking in the TEM10 mode. I proposed that the alignment might be sketchy. I've been fiddling with the MC alignment sliders for the last hour and a half or so, but I think I'm not 100% in tune with the 3 mirror parameter space. The mode cleaner now locks, but I'm not in love with its' alignment. The WFS are definitely catywhompus. Before doing hardware things like recentering the WFS, I'm going to wait until tomorrow to consult with an alignment expert.
In case this is helpful for tomorrow, before I touched any of the sliders:
Optic, Pitch, Yaw
MC1, 3.1459, -0.7200
MC3, -0.8168, -3.0700
MC2, 3.6360, -1.0576
Now that mode cleaner locks, although not in a great alignment:
MC1, 3.1089, -0.7320
MC3, -0.7508, -3.0770
MC2, 3.6610, -1.0786
If I knew how to kill my script to unlock the mode cleaner, I would. But I sourced it, and Rob didn't know earlier this evening how to kill something which is started with 'source' since it doesn't seem to get a process number like when you './' to run a script. So the Mode Cleaner will probably be unlocked in the morning, and it may be persnickity to get it relocked, especially if the tree people are doing tree things with giant trucks again in the morning.
So that I can collect a bit of free-swinging Mode Cleaner data, I started a script to wait 14400 seconds (4 hours), then unlock the mode cleaner. It should unlock the MC around 4am. As soon as someone gets in in the morning, you can relock it. I should have plenty of data by then.
Today I set up the EUCLID long range michelson design on the SP table; It's the same as the setup posted earlier, but without the pickoff (at PD1), which can be added later, and a few other minor changes (moved lenses, mirrors, PDs - nothing major). I hooked up the two PD's to the oscilliscope and got a readout that pointed to more power hitting PD2 than PD3.
In the last hour or so the elog crashed. I have restarted it.
Yesterday, Alex attached the old frame builder 1.5 TB raid array to linux1, and tested to make sure it would work on linux1.
This morning he tried to start a copy of the current /cvs/cds structure, however realized at the rate it was going it would take it roughly 5 hours, so he stopped.
Currently, it is planned to perform this copy on this coming Friday morning.
I measured the magnitude of modulation as a function of frequency using the optical spectrum analyzer and an oscilloscope while generating signals using a Marconi signal generator; the results are shown in the attached plot and are compared to the expected modulation given the measured transfer function of the circuit and the nominal modulation index of the EOM used (13 mrad/V). Using the oscilloscope, I found the resonant peaks to be at 11.11 MHz, 29.57 MHz, and 54.70 MHz. There are several different colors on the plot; this is because I had to take the data in several different segments and had to switch to measuring a different sideband partway through the measurment. I also separately found the modulation at each resonant peak for each sideband. The magnitude of modulation was measured by finding the ratio between the magnitude of the carrier and sideband powers using an oscilloscope, and calculating the magnitude of modulation from this. This method was also used to quantify the dependence of modulation magnitude on input power at each resonant peak; these results are also attached. These same results can also be plotted as modulation magnitude as a function of voltage into the resonant circuit; this is also attached (I'm not sure which is more useful).
In order to produce these results (get the measurements in mrad/V) it was necessary to measure the gain of the amplifier. I used the signal generator to input signals of varying power and measured the output signal voltage using the oscilloscope; I then repeated this process at each resonant frequency. From this I was able to calculate the gain of the amplifier to be 28.1 dB at 11.11 MHz, 27.4 dB at 29.57 MHz, and 25.7 dB at 54.70 MHz. These values are in the same ballpark as the values in the Mini Circuits data sheet (all values are ~25-28 MHz).
Looks like someone rebooted nodus at ~3 PM today but did not elog it. Also the SVN is not running. Why?
The Nodus business was me....my bad. Nodus and the elog were both having a bad day (we couldn't ssh into nodus from op440m (which doesn't depend on the names server)), so I called Alex, and he fixed things, although I think that all he did was reboot. I then restarted the elog per the instructions on the wiki.
Spent a lot of time aligning tonight. The BS is not staying put--sometimes after a lock loss it gets badly mis-aligned.
DD handoff is working, after putting beam on REFL diodes and running senseDRM script.
Rich Abbott, Rana
Summary: We found that the 3mm InGaAs photodiodes from eGTRAN which are being used for the DC Readout in eLIGO are bad. The QE is ~50%. We will have to replace them ASAP.
Valera and Nic Smith have pointed out out a factor of ~2 discrepancy between the estimated power transmission to the dark port in H1 and L1. So we decided to measure the QE of the accused diodes.
The data of the QE and dark current are attached here.
We used a 1064 nm CrystaLaser (which does not have a very stable power output). We attenuated the light with an ND1.0 for all measurements.
The photocurrent is estimated by reading out the voltage across one leg of the differential drive of the DC PD preamp. The photocurrent goes across a 100 Ohm resistor and then through 2 gain of 1 stages to get to this testpoint, so the overall transimpedance gain is 100 Ohms for this measurement.
By far, the Ophir power meter is the biggest source of error. Its absolute calibration is only 5% and the variation across the sensor face is ~5%. There are some hot and not hot spots on the face which can make even more variation, but we tried to avoid these.
We also inserted the power meter very close to the time when we read the voltage, so that the photocurrent and power estimates are made within 10 seconds of each other. This should reduce the error from the laser's power fluctuations.
All diodes still had the glass case on. We measured the reflected power to be ~5-7% of the incident power. This reflected power is NOT accounted for in these estimates.
Punch line: The eGTRAN diodes that we currently use are definitely bad. The JDSU and EG&G 2mm diodes have a better QE. We should immediately purchase 3 mm versions and get them cut and measured to be ready for the Sep. 1 commissioning surge.
I was able to observe the three sets of modulation sidebands created by the EOM + triply resonant circuit yesterday. Quantitative results will be posted later.
While writing my progress report, I redrew the Guralp breakout box circuit diagram with all the changes marked. Since only one hard copy exists, I thought it might be useful to post my drawing up in case it is needed for any reason. The two drawings are the same - the second has just been broken into two parts to make it easier to fit on a normal 8.5 x 11 or A4 sheet of paper. The gains for each opamp have not been marked, but they could very easily be added in if necessary. The black resistances and capacitances are the originals. All changes have been indicated in blue.
the servo needs some work.
2 day trend
This afternoon we tried to improve the mode matching of the beam to the PMC. To do that we tuned the positions of the two lenses on the PSL table that come before the PMC.
We moved the first lens back an forth the without noticing any improvement on the PMC transmitted and reflected power. Then we moved the first backwards by about one cm (the order is set according to how the beam propagates). That made the things worse so we moved also the second lens in the same direction so that the distance in between the two didn't change significantly. After that, and some more adjustments on the steering mirrors all we could gain was about 0.2V on the PMC transmission.
We suspect that after the problems with the laser chiller of two months ago, the beam size changed and so the mode matching optics is not adequate anymore.
We have to replace the mode matching lenses with other ones.
The second set of Guralp channels is now plugged into the PEM ADCU, into channels which are confirmed to be working. (Method: 1Vpp sine wave into channel, check with DataViewer).
Direction, Channel Name, .ini chnum, BNC plug # on ADCU
Vertical: C1:PEM-SEIS_GUR_VERT, 15023, #24
N/S (should be Y when the seismometer is put in place): C1:PEM-TEMP_2, 15001, #2
E/W (should be X when the seismometer is put in place): C1:PEM-TEMP_3, 15002, #3
There is IFO work going on, so I don't want to rename the channels / restart fb40m until a little later, so I'll just use the old TEMP channel names for now.
There is something totally wrong with the E/W channel. I can look at all 3 channels on a 'scope (while it's on battery, so the op-amps in the breakout box aren't grounded), and VERT and NS look fine, and when I jump around ("seismic testing"), they show spikes. But the EW channel's signal on the 'scope is way smaller, and it doesn't show anything when I jump.
I might use the handheld Guralp tester breakout box to check the seismometer. Also, a suspicion I have is that whoever put the box back in on Friday night after our final noise measurements left the inputs shorted for this one channel. It's the 3rd channel in the set, so it would be most likely to be stuck shorted... Investigations will ensue.
All the channels are now good, and all the names are back to making sense.
The problem with EW2 was in fact that the alligator clip used to short the inputs during the noise test Friday night was left in the box. Not great, but now it's taken care of, and we have recorded data of the noise of the breakout box, so we can include that in our plots to see if we're at the limit of how good we can do at subtracting noise.
The channels are now named thusly:
C1:PEM-SEIS_GUR_VERT (BNC input #24, .ini channel #15023)
C1:PEM-SEIS_GUR_EW (BNC input #3, .ini channel #15002)
C1:PEM-SEIS_GUR_NS (BNC input #2, .ini channel #15001)
C1:PEM-SEIS_MC1_X (BNC input #11, .ini channel #15010)
C1:PEM-SEIS_MC1_Y (BNC input #12, .ini channel #15011)
C1:PEM-SEIS_MC1_Z (BNC input #10, .ini channel #15009)
C1:PEM-SEIS_MC2_Y (Ranger, which for the Huddle Test is oriented VERTICALLY) (BNC input #4, .ini channel #15003)
Now we wait.....and tomorrow extract the noise of each of the seismometers from this!
We discussed a preliminary game plan for this project. The thing I really want to see is an ETMX RCG controller hooked into the existing frontend via reflective memory, and the 40 m behaving normally with this hybrid system, and my list is geared toward this. I suspect the list may cause controversy.
+ copy the MDC filters into SAM, and make sure everything looks good there with DTT and SR785.
+ get interface / wiring boards from Wilson House, to go between megatron and the analog ETMX system
+ test tying the ETMX pendulum and bare-bones SAM together (use existing watchdogs, and "bare-bones" needs defining)
+ work some reflective memory magic and create the hybrid frontend
In parallel with the above, the following should also happen:
+ MEDM screen design
+ add non-linear bits to the ETMX MDP/MDC model system
+ make game plan for the rest of the RCG frontend
I've added the PIT and YAW dofs to the MDC and MDP systems. The pendula frequencies in MDP are 0.8, 0.5, 0.6 Hz for POS, PIT, and YAW respectively. The three dofs are linear and uncoupled, and stable, but there is no modeled noise in the system (yet) and some gains may need bumping up in the presence of noise. The MDC filters are identical for each dof (3:0.0 and Cheby). The PIT and YAW transfer functions look pretty much like the one Rana recently took of POS, but of course with the different pendulum frequencies. I've attached one for YAW.
These are the settings which determine the transmon (eg, TRX) amplitude, and which are updated by the matchTransMon scripts.
For the X arm
op440m:AutoDither>tdsread C1:LSC-TRX_GAIN C1:LSC-LA_PARAM_FLT_01 C1:LSC-LA_PARAM_FLT_00
For the Y arm
op440m:AutoDither>tdsread C1:LSC-TRY_GAIN C1:LSC-LA_PARAM_FLT_04 C1:LSC-LA_PARAM_FLT_03
With the high power meter I measured the reflected power when the PMC was unlocked and used that to obtain the calibration of the PMC-REFL PD: 1.12V/W.
P_in = 1.98W ; P_trans = 1.28W ; P_refl = 0.45W
From that I estimated that the losses account to 13% of the input power.
I checked both the new and the old elogs to see if such a measurement had ever been done but it doesn't seems so. I don't know if such a value for the visibility is "normal". It seems a little low. For instance, as a comparison, the MC visibility, is equal to a few percents.
Also Rana measured the transmitted power after locking the PMC on the TEM20-02: the photodiode on the MEDM screen read 0.325V. That means that a lot of power is going to that mode.
That makes us think that we're dealing with a mode matching problem with the PMC.
I aligned the MZ. The reflection went from .86 to .374
I think the MZ pzt is broken/failing. I'm not sure how else to explain this behavior.
The first bit of the time series is a triangle wave into the DC offset (output) field, over approximately the whole range (0-250V). You can see the fringe visbility is quite small. The triangle wave is stopped, and I then maxed out the offset slider to get to the "high" power point from the triangle wave sweep. Then for a little while with the PZT is held still, and the power still increases. The MZ is then locked, and you can see the PZT voltage stay about the same but the power continues to rise over the next ~10 minutes or so.
This plot answers the previous question, and raises a new one--what the heck is MZTRANSPD? I'd guess the pins are unconnected--it's just floating, and somehow picking up the MZ_PZT signal.
The offending beam dump has been removed, and the PMC relocked.
Maybe it was Russell Crowe
This is very nice. We have, for the first time, a real time plant with which we can test our changes of the control system. From my understanding, we have a control system with the usual POS/PIT/YAW matrices and filter banks. The outputs go to a separate real-time system which is running something similar and where we have loaded the pendulum TF as a filter. Cross-couplings, AA & AI filters, and saturations to come later.
The attached plot is just the same as what Peter posted earlier, but with more resolution. I drove at the input to the SUSPOS filter bank and measured the open loop with the loop closed. The loop wants an overall gain of -0.003 or so to be stable.