Koji was correct.
When you estimate the variance of the population, you have to use unbiased variance (not sample variance). So, the estimate to dx in the equations Koji wrote is
dx = sqrt(sum(xi-xavg)/(n-1))
It is interesting because when n=2, statistical error of the averaged value will be the same as the standard deviation.
dXavg = dx/sqrt(n) = stdev/sqrt(n-1)
In most cases, I think you don't need 10 % precision for statistical error estimation (you should better do correlation analysis if you want to go further). You can simply use dx = stdev if n is sufficiently large (n > 6 from plot below).
Makes sense. I mixed up n and n-1
Probability function: X = (x1 + x2 + ... + xn)/n, where xi = xavg +/- dx
Xavg = xavg*n/n = xavg
dXavg^2 = n*dx^2/n^2
=> dXavg = dx/sqrt(n)
Xavg +/- dXavg = xavg +/- dx/sqrt(n)
On the day before yesterday and in this morning, I measured loss map of ETMX. I reported the method I used to change the beam spot on ETMX below.
Round trip loss was measured for 5 x 5 points. The result is below.
455.4 +/- 21.1 437.1 +/- 21.8 482.3 +/- 21.8 461.6 +/- 22.5 507.9 +/- 20.1
448.4 +/- 20.7 457.3 +/- 21.2 495.6 +/- 20.2 483.1 +/- 20.8 472.2 +/- 19.8
436.9 +/- 19.3 444.6 +/- 19.7 483.0 +/- 19.5 474.9 +/- 20.9 498.3 +/- 18.7
454.4 +/- 18.7 474.4 +/- 20.6 487.7 +/- 21.4 482.6 +/- 20.7 487.0 +/- 19.9
443.7 +/- 18.6 469.9 +/- 20.2 482.8 +/- 18.7 480.9 +/- 19.5 486.1 +/- 19.2
The correspondence between the loss shown above and the beam spot on ETMX is shown in the attached figure. In the figure, "up" and "right" indicate direction of shift of the beam spot when you watch it via the camera (ex. 455.4 ppm corresponds to the highest and rightest point in the view via the camera).
This result is consistent withe previous result of 561.19 +/- 14.57 ppm ericq got with ASDC and reported in elog 10248 if the discussion I reported in 11819 is taken into account. Elog 11819 says in short that the strange behavior of ASDC could give us 60-70 ppm error.
The reason why the error is larger than that of the measurement for ETMY is that the noise of POX is larger than that of POY. But I am not sure to what extent the statistical error needs to be reduced.
How I shifted the beam spot on ETMX:
Basically, the method was same as one used for Y arm. Different point is: for Y arm we have two steering mirrors TT1&2, but for X arm we have only one steering mirror BS. Then in order to shift incident beam so that the beam spot on ITMX does not change, I ran the dithering of X arm as well as that of Y arm and added offsets to both dither loops that caused same amount of shift on ETMX and ETMX. Thanks to the symmetry between X arm and Y arm, the dithering of Y arm ensured that the beam spot on ITMX was unchanged as well as that of ITMY. The idea of this method is schematically shown in Attachment 2.
The calibration of how much the beam spot shifted is based on the results of elog 11846 . The offset was [-15,-7.5,0,7.5,15]x[-5,-2.5,0,2.5,5] for pitch and yaw, respectively.
I measured round trip loss of Y arm. The alignment of relevant mirrors was set ideal with dithering (no offset).
round trip loss of Y arm: 166.2 +/- 9.3 ppm
(In the error, only statistic error is included.)
How I measured it:
I compared the power of light reflected by Y arm (measured at AS) when the arm was locked (P_L) and when ETMY was misaligned (P_M). P_L and P_M can be described as
The reason why P_L takes this form is: (1-alpha)*4T_ITM/(T_tot)^2 is intracavity power and then product of intracavity power and loss describes the power of light that is not reflected back. Here, alpha is power ratio of light that does not resonate in the arm (power of mismatched mode and modulated sideband), and T_tot is T_ITM+T_loss. Transmissivity of ETM is included in T_loss. I assumed alpha = 7%(mode mismatch) + 2 % (modulation) (elog 11745)
After some calculation we get
Here, higher order terms of T_ITM and (T_loss/T_ITM) are ignored. Then we get
Using this formula, I calculated T_loss. P_L and P_M were measured 100 times (each measurement consisted of 1.5 sec ave.) each and I took average of them. T_ETM =13.7 ppm is used.
-- This value is not so different from the value ericq reported in July (elog 10248).
-- This method of measuring arm loss is NOT sensitive to T_ITM. In contrast, the method in which loss is obtained from finesse (for example, elog 11740) is sensitive to T_ITM.
In the method I'm now reporting,
but in the method with finesse,
In the latter case, if relative error of T_ITM is 10%, error of T_loss would be 1000 ppm.
So it would be better to use power of reflected light when you want to measure arm loss.
Due to the strange behavior (elog 11815) of ASDC level, we checked if it is possible to use POYDC instead of ASDC to measure the power of reflected light of YARM. Attached below is the spectrum of them when the arm is locked. This spectrum shows that it is not bad to use POYDC, in terms of noise. The spectrum of them when ETMY is misaligned looked similar.
So I am going to use POYDC instead of ASDC to measure arm loss of YARM.
Ed by KA:
The spectra of POYDC and ASDC were measured. We foudn that they have coherence at around 1Hz (good).
It told us that POYDC is about 1/50 smaller than ASDC. Therefore in the attached plot, POYDC x50 is shown.
That's the meaning of the vertical axis unit "ASDC".
Tonight I measured "loss map" of ETMY. The method to calculate round trip loss is same as written in elog 11810, except that I used POYDC instead of ASDC this time.
How I changed beam spot on ETMY is: elog 11779.
I measured round trip loss for 5 x 5 points. The result is below.
494.9 +/- 7.6 356.8 +/- 6.0 253.9 +/- 7.9 250.3 +/- 8.2 290.6 +/- 5.1
215.7 +/- 4.8 225.6 +/- 5.7 235.1 +/- 7.0 284.4 +/- 5.4 294.7 +/- 4.5
205.2 +/- 6.0 227.9 +/- 5.8 229.4 +/- 7.2 280.5 +/- 6.3 320.9 +/- 4.3
227.9 +/- 5.7 230.5 +/- 5.5 262.1 +/- 5.9 315.3 +/- 4.7 346.8 +/- 4.2
239.7 +/- 4.5 260.7 +/- 5.3 281.2 +/- 5.8 333.7 +/- 5.0 373.8 +/- 4.9
The correspondence between the loss shown above and the beam spot on ETMY is shown in the following figure. In the figure, "downward" and "left" indicate direction of shift of the beam spot when you watch it via the camera (ex. 494.9 ppm corresponds to the lowest and rightest point).
Edited below on 28th Nov.
To shift the beam spot on ETMY, I added offset in YARM dither loop. The offset was [-30,-15,0,15,30]x[-10,-5,0,5,10] for pitch and yaw, respectively. How I calibrated the beam spot is basically based on elog 11779, but I multiplied 5.3922 for vertical direction and 4.6205 for horizontal direction which I had obtained by caliblation of oplev (elog 11785).
Edited above on 28 th Nov.
I will report the detail later.
Here, I upload data I took last night, including the power of reflected power (locked/misaligned) and transmitted power for each point (attachement 1).
And I would like to write about possible reason why the loss I measured with POYDC and the loss I measured with ASDC are different by about 60 - 70 ppm (elog 11810 and 11818). The conclusion I have reached is:
It could be due to the strange bahavior of ASDC level.
This difference corresponds to the error of ~2% in the value of P_L/P_M. As reported in elog 11815, ASDC level changes when angle of the light reflected by ITMY changes, and 2% change of ASDC level corresponds to 10 urad change of the angle of the light according to my rough estimation with the figure shown in elog 11815 and attachment 2. This means that 2% error in P_L/P_M could occur if the angle of the light incident to YARM and that of resonant light in YARM differ by 10 urad. Since the waist width of the beam is ~3 mm, with the 10 urad difference, the ratio of the power of TEM10 mode is , where . This value is reasonable; in elog 11743 Gautam reported that the ratio of the power of TEM10 was ~ 0.03, from the result of cavity scan. Therefore it is possible that the angle of the light incident to YARM and that of resonant light in YARM differ by 10 urad and this difference causes the error of ~2% in P_L/P_M, which could exlain the 60 - 70 ppm difference.
The rsync job to sync our frames over to the cluster has been on a 20 MB/s BW limit for awhile now.
Dan Kozak has now set up a cronjob to do this at 10 min after the hour, every hour. Let's see how this goes.
You can find the script and its logfile name by doing 'crontab -l' on nodus.
Still seems to be running without causing FB issues. One thought is that we could look through the FB status channel trends and see if there is some excess of FB problems at 10 min after the hour to see if its causing problems.
I also looked into our minute trend situation. Looks like the files are comrpessed and have checksum enabled. The size changes sometimes, but its roughly 35 MB per hour. So 840 MB per day.
According to the wiper.pl script, its trying to keep the minute-trend directory to below some fixed fraction of the total /frames disk. The comment in the scripts says 0.005%,
but I'm dubious since that's only 13TB*5e-5 = 600 MB, and that would only keep us for a day. Maybe the comment should read 0.5% instead...
Still seems to be running without causing FB issues.
I'm not so sure. I just was experiencing some severe network latency / EPICS channel freezes that was alleviated by killing the rsync job on nodus. It started a few minutes after ten past the hour, when the rysnc job started.
Unrelated to this, for some odd reason, there is some weirdness going on with ssh'ing to martian machines from the control room computers. I.e. on pianosa, ssh nodus fails with a failure to resolve hostaname message, but ssh nodus.martian succeeds.
So today, after an "rm" error while working with the autoburt.pl script and burt restores in general, I asked Dan Kozak how to actually look at the backup data. He said there's no way to actually look at it at the moment. You can reverse the rsync command or ask him to grab the data/file if you know what you want. However, in the course of this, we realized there was no /cvs/cds data backup.
Turns out, the rsync command line in the script had a "-n" option. This means do a dry run. Everything *but* the actual final copying.
I have removed the -n from the script and started it on nodus, so we're backing up as of 5:22pm today.
I'm thinking we should have a better way of viewing the backup data, so I may ask Dan and Stewart about a better setup where we can login and actually look at the backed up files.
In addition, tomorrow I'm planning to add cron jobs which will put changes to files in the /chans and /scripts directories into the SVN on a daily basis, since the backup procedure doesn't really provide a history for those, just a 1 day back backup.
We can't compile any changes to the LSC or the GCV models since Jamie's new script / program isn't found. I don't know where it is (I can't find it either), so I can't do the compiling by hand, or point explicitly to the script. The old way of compiling models in the wiki is obsolete, and didn't work :(
Sorry about that. I had modified the path environment that pointed to the rtcds util. The rtcds util is now in /opt/rtcds/caltech/c1/scripts/rtcds, which is in the path. Starting a new shell should make it available again.
Added TRX and TRY and POY11_I_ERR and POX11_I_ERR to the c1lsc.mdl using a new-style DAQ Channels block, recompiled, installed, started the model, all good. Restarted the daqd on the framebuilder, and everything is green. I can go back and get recorded data using dataviewer (for the last few minutes since I started fb), so it all looks good.
Note on the new DAQ Channels block: Put the text block (from CDS_PARTS) at the same level as the channel you want to save, and name it exactly as it is in the model. The code-generator will add the _DQ for you. i.e. if you define a channel "TRY_OUT_DQ" in the lsc model, you'll end up with a channel "C1:LSC-TRY_OUT_DQ_DQ".
This means we can't (a) record TRY or (b) add the Q quadrature of the beat PD to the real time system tonight.
We're going to try just using Yuta's pynds script to capture data in real time, so we can keep working for tonight.
Finally I found a company who can do Koji's improved -hard to make- specification on ruby or sapphire wire standoff.
NOT POLISHED excimer laser cut, wire groove radius R 0.0005" + - 0.0002"
$250 ea at 50 pieces of order
Atm 1 & 5, showing the ruby R ~10 mm as it is seated on Al SOS test mass
Atm. 2, 3 & 4 chipped long edges with SOS sus wire OD 43 micron as calibration
Ruby wire standoff received from China. I looked one of them with our small USB camera. They did a good job. The long edges of the prism are chipped.
The v-groove cutter must avoid them. Pictures will follow.
Bluebean Optical Tech Limited of Shanghai delivered 50 pieces red ruby prisms with radius. The first prism pictures were taken at June 5
and it was retaken again as BB#1 later
More samples were selected randomly as one from each bag of 5 and labeled as BB#2.......6
The R10 mm radius can be seen agains the ruler edge. The v-groove edge was labeled with blue marker and pictures were taken
from both side of this ridge. The top view is shown as the wire laying across on it.
SOS sus wire of 43 micron OD used as calibration as it was placed close to the side that it was focused on.
The V-groove ridge surface quality was evaluated based on as scale of 1 – 10 with 10 being the most positive.
Remaining thing to examin, take picture of the contacting ridge to SOS from the side.
3-4 hrs ago we run out of nitrogen. We are back to Vacuum Normal
I have attached the result of running the PID script on the seismometer with the can on. The daily fluctuations are no more than 0.07 degrees off from the setpoint of 39 degrees. Not really sure what happened in the past day to cause the strange behavior. It seems to have returned back to normal today.
I'm running a comsol job on optimus in a tmux session named cryocavs. Should be done in less than 24 hours, judging by past durations.
I'm temporarily writing frames into a tempfs, which is a filesystem that exists purely in memory. There should be ZERO IO contention for this filesystem, so if the daqd failures are due to IO then all problems should disappear. If they don't, then we're dealing with some other problem.
There will be no data saved during this period.
I have reverted daqd to the previous configuration, so that it's writing frames to disk. It's still showing instability.
In order to figure out what downsampling ratio we can take, we need to determine the running time of the fxlms_filter() function. If the filter length is equal to 5000, downsampling ratio is equal to 1, number of witness channels is 1 then with ordinary compilation without speed optimization one call runs for 0.054 ms (milli seconds). The test was done on the 3 GHz Intel processor. With speed optimization flags the situation is better
-O1 0.014 ms, -O3 0.012 ms
However, Alex said that speed optimization is not supported at RCG because it produce unstable execution time for some reason. However, by default the kernel should optimize for size -Os. With this flag the running time is also 0.012 ms. We should check if the front-end machine compilers indeed use -Os flag during the compilation and also play with speed optimization flags. Flags -O3 and -Os together might also give some speed improvement.
But for now we have time value = 0.012 ms as running time for 5000 coefficient filter, 1 witness channel and downsample ratio = 1. Now, we need to check how this time is scaled if we change the parameters.
5000 cofficients - 0.012 ms
10000 coefficients - 0.024 ms
15000 coefficients - 0.036 ms
20000 coefficients - 0.048 ms
We can see that filter length scaling is linear. Now we check downsampling ratio
ratio=1 - 0.048 ms
ratio=2 - 0.024 ms
ratio=4 - 0.012 ms
Running time on the dependance of downsample ratio is also linear as well as on the dependence of the number of witness channels and degrees of freedom.
If we want to filter 8 DOF with approximately 10 witness channels for each DOF, then 5000 length filter will make 1 cycle for ~1 ms, that is good enough to make the downsample ratio equal to 4.
Things get a little bit complicated when the code is called for the first time. Some time is taken to initialize variables and check input data. As a result the running time of the first cycle is ~0.1 ms for 1 DOF that is ~10 times more then running time of an ordinary cycle. This effect takes place at this moment when one presses reset button in the c1oaf model - the filter becomes suspended for a while. To solve this problem the initialization should be divided by several (~10) parts.
This is long overdue, but our burt files for SDF now live in the LIGO userapps SVN, as they should.
The canonical files live in places like /opt/rtcds/userapps/release/cds/c1/burtfiles/c1x01_safe.snap and are symlinked to files like /opt/rtcds/caltech/c1/target/c1x01/c1x01epics/burt/safe.snap
The 40m safety audit will be at Wednesday afternoon, March 3
Please participate getting our lab to inspection grade safety level.
Correction list by visiting safety committee, Haick Issaian is not shown:
1, update laser, crane operator list and post it
2, check fire extinguishers monthly, date and initials must be on the tags
3, move drinking water tower so it does not block fire extinguisher
4, post updated crane doc at cranes
5, post present phone lists at IFO room phones
6, emergency laser shutoff at the south end must be mounted without C-clamp
7, use heavy cable tie to insure position of mag-fan on cabinet top
Additional to do list:
a, safety glasses to be cleaned
b, let the electrical shop fix Rack-AC power to optical tables at the ends
c, measure transmission of laser safety glasses
d, update IFO outside door info signs
e, update laser inventory and post it
f, schedule annual crane inspection and renew maintenance contract
g, PSL enclosure inner shelf needs a good clean up so it is earthquake safe
Completed with the exception of d and g
Recommended correction list:
1, refill- upgrade first aid boxes
2, maintain 18" ceiling to bookshelf clearance so the ceiling fire sprinklers are not blocked: room 101
3, label chilled water supply & return valves in IFO room
4, calibrate bake room hoods annually
5, update safety sign at fenced storage
40m still to do list:
1, clean and measure all safety glasses
2, annual crane inspection is scheduled for 8am March 19, 1013
3, make PSL encloser shelf earthquake proof
Do you see something that is not safe? Add it to this list please.
We had our annual safety inspection today. Our SOPs are outdated. The full list of needed correction will be posted tomorrow.
The most useful found was that the ITMX-ISCT ac power is coming from 1Y1 rack. This should actually go to 1Y2 LSC rack ?
Please test this so we do not create more ground loops.
Annual crane inspection is scheduled for 8-11am Monday, March 17, 2014
The control room Smart UPS has two red extension cords that has to be removed: Nodus and Linux1
Last long extension cord removed from 1Y1 to ITMX-ISCT
The AC power strip at ITMX-ISCT is coming from wall L#26
Be aware that this may affect POP QPD and POP RF Thorlabs PD
Late adition: CHECK all viewport covers.
A, transparent Lexan sheet is protecting glass windows in a horizontal position
B, metal housing protection is required on each viewport except signal ports
C, signal ports should be shielded by optical table enclosure
We have to cover this window-camera with implosion proof cover or just remove it and blank it.
Question number 2: Do our vertically positioned windows with flip able covers require protective lexan ? NO 5-5-2014
Safety audit went soothly. We thank all participients.
1, Bathroom water heater cable to be stress releived and connector replaced by twister lock type.
2, Floor cable bridge at the vacuum rack to be replaced. It is cracked.
3, Sprinkler head to be moved eastward 2 ft in room 101
4, Annual crane inspection is scheduled for 8am Marc 3, 2015
5, Annual safety glasses cleaning and transmission measurement will get done tomorrow morning.
Safety glasses were measured and they are all good. I'd like to measure your personal glass if it is not on this picture.
Safety audit went smothly.
Crane inspection is scheduled for March 4
Safety glasses will be measured before April 1
Bob cleaned the safety glasses. They were sonicated in warm 2% Liquinox water for 10 minutes. Steve checked them by transmission measurement of 1064 nm at 150 mW
Linus-1, Nodus and others ac cords can be moved over to new blank yellow extension cord with multiple recepticals.
Remove two red extension cords going to Smart UPS
Emergency exit lights were inspected: 2 out of 13 batteries have to be replaced
One of the Halon fire extinguishers needs to be recharged out of 8
Please do participate in preparation for the upcoming safety audit on Feb 28
Batteries replaced and cylinder recharged. Please clean up your experimental set up if it is blocking breakers or entry way etc.
I will start the final clean up 2pm today.
1064 nm transmison were measured of 40m safety glasses as shown . Their performance did not degrade. They are as good as their labels.
Safety glasses 1064 nm transmission measured at ~200 mW level. They are all good.
This is the third safety glasses that I found laying around in the IFO lab lately. Safety glasses must be worn ALL times in the IFO room!
This rule is essential for your protection! Please do not enter if you can not put up with this regulation!
Lightwave M126-1064-700 lasers sn415 at east end of the Y arm and sn201 at the AP table are connected individually to one each EMERGENCY LASER SHUT OFF SWITCH.
The PSL had one 1064 nm beam to be blocked around the north east side. The end enclosures are fine.