[Koji / Suresh]
We worked on the 1064 beating a bit more.
- First of all, FSS and FSS SLOW servo were disabled not to have variating Xtal temp for the PSL.
- The PSL Xtal temp (T_PSL) was scanned from 22deg-45deg while we search the Xtal temp (T_Xend) for the Xend laser to have the beat freq well low (f<30MHz).
The pumping current for each laser was I_PSL = 2.101 [A] and I_Xend = 2.000 [A]
For a certain T_PSL, we found multiple T_Xend because the freq of the laser is not a monotonic function of the Xtal temperature. (see the innolight manual).
T_Xend to give us the beating was categorized in the three sets as shown in the figure. The set on "curve2" is the steadiest one. (Use this!)
The trends were quite linear but the slope was slightly off from the unity.
- T_PSL was scanned to see the trend of the PMC output.
The PMC was sometimes locked to the mode with lower transmission (V_PMCT ~ 3.0V).
When T_PSL ~ 31deg we consistently locked the PMC at higer transmission (V_PMCT ~ 5.3V).
At the moment we decided the operating point of T_PSL = 32.25 deg, V_PMCT = 5.34, where we found the beat at T_Xend=38.28deg.
- We cleaned up the PSL table more than how it was. Returned the tools to their original places.
The X-end laser was shut down and was left on the PSL table.
Kiwamu can move the X-end laser to the Xend and realign it.
Then we should be able to see the green beating quite easily.
Thesedays we were continuously annoyed by unELOGGED activities of the interferometer.
MC2 LOCKIN was left on and has continuously injected frequency noise and beam pointing modulation
during all of the comissioning / vent preparation.
C1:SUS-MC2_LOCKIN2_OSC_FREQ was 0.075
C1:SUS-MC2_LOCKIN2_OSC_CLKGAIN was 99
For more than a week ago we noticed that the curve of the MC WFS stripchart suddenly got THICKER.
MC WFS, arm transmission, beam pointing... everything was modulated.
It was not WFS instability, and it was not the cavity mirrors.
Today I made the investigation and finally tracked down the cause of this issue to be on MC2 suspension.
Then it was found that this LOCKIN was ON.
There is no direct record of this lockin in the frame files.
From the recorded channel "C1:IOO-WFS2-YAW_OUT16" (which is the trace on the StripTool chart on the wall)
It was turned on at July 10th, 2:00UTC (July 9th, 7PM PDT)
Yes, this was not ELOG'd by me, unfortunately. This was the MC tickler which I described to some people in the control room when I turned it on.
As Koji points out, with the MCL path turned off this injects frequency noise and pointing fluctuations into the MC. With the MCL path back on it would have very small effect. After the pumpdown we can turn it back on and have it disabled after lock is acquired. Unfortunately, our LOCKIN modules don't have a ramp available for the excitation and so this will produce some transients (or perhaps we can ezcastep it for now). Eventually, we will modify this CDS part so that we can ramp the sine wave.
I've written a new TICKLE script using the newly found 'cavget' and 'cavput' programs. They are in the standard epics distribution as extension binaries. They allow multichannel read/write as well as ramping, delays, incremental steps, etc. http://www.aps.anl.gov/epics/tech-talk/2012/msg01465.php.
Running from the command line, they seem to work fine, but I've left it OFF for now. I'll switch it into the MC autolocker at some point soon.
I found the PSL laser has been off for four hours. Nobody seemed to know why.
I just turned it on and it is now providing about 10% lower power compared with one before the shutdown.
Let's keep the eyes on the power if it can recover as the housing gets warm.
Frame builder is down. PRM has tripped its watch dogs. I have reset the watch dog on PRM and turned on the OPLEV. It has damped down. Unable to check what happened since FB is not responding.
There was an minor earthquake yesterday morning which people could feel a few blocks away. It could have caused the the PRM to unlock.
Jamie,Rolf, is it okay or us to restart the FB?
If it's down it's alway ok to restart it. If it doesn't respond or immediately crashes again after restart then it might require some investigation, but it should always be ok to restart it.
I tried restarting the fb in two different ways. Neither of them re-established the connection to dtt or epics.
1) I restarted the fb from the control room console with the 'shutdown' command. No change.
2) I halted the machine with 'shutdown -h now' and restarted it with the hardware reset button on its front-panel. No change.
The console connected to the fb showed that the network file systems did not load. Could this have resulted in failure to start several services since it could not find the files which are stored on the network file system?
The fb is otherwise healthy since I am able to ssh into it and browse the directory structure.
The fb is okay. Rana found that it works on Pianosa, but not on Allegra or Rossa. It also works on Rosalba, on which Jamie recently installed Ubuntu.
The white fields on the medm 'Status' screen for fb are an unrelated problem.
Please be conscious of what components are doing what. The problem you were experiencing was not "frame builder down". It was "dtt not able to connect to frame builder". Those are potentially completely different things. If the front end status screens show that the frame builder is fine, then it's probably not the frame builder.
Also "epics" has nothing whatsoever to do with any of this. That's a completely different set of stuff, unrelated to DTT or the frame builder.
I think the daqd process isn't running on the frame builder.
I tried telnetting' to fb's port 8087 (telnet fb 8087) and typing "shutdown", but so far that is hanging and hasn't returned a prompt to me in the last few minutes. Also, if I do a "ps -ef | grep daqd" in another terminal, it hangs.
I wasn't sure if this was an ntp problem (although that has been indicated in the past by 1 red block, not 2 red blocks and a white one), so I did "sudo /etc/init.d/ntp-client restart", but that didn't make any change. I also did an mxstream restart just in case, but that didn't help either.
I can ssh to the frame builder, but I can't do another telnet (the first one is still hung). I get an error "telnet: Unable to connect to remote host: Invalid argument"
Thoughts and suggestions are welcome!
CPU load seems extremely high. You need to reboot it, I think
controls@fb /proc 0$ cat loadavg
36.85 30.52 22.66 1/163 19295
This CPU load may have been me deleting some old frame files, to see if that would allow daqd to come back to life.
Daqd was segfaulting, and behaving in a manner similar to what is described here: (stack exchange link). However, I couldn't kill or revive daqd, so I rebooted the FB.
Things seem ok for now...
[Joe, Jamie, Alex]
I asked Alex which cron to use (dcron? frcron?). He promptly did the following:
rc-update add dcron default
Copied the wiper.pl script from LLO to /opt/rtcds/caltech/c1/target/fb/
At that point, I modified wiper.pl script to reduce to 95% instead of 99.7%.
I added controls to the cron group on fb:
sudo gpasswd -a controls cron
I then added the wiper.pl to the crontab as the following line using crontab -e.
0 6 * * * /opt/rtcds/caltech/c1/target/fb/wiper.pl --delete &> /opt/rtcds/caltech/c1/target/fb/wiper.log
Note, placing backups on the /frames raid array will break this script, because it compares the amount in the /frames/full/, /frames/trends/minutes, and /frames/trends/seconds to the total capacity.
Apparently, we had backups from September 27th, 2010 and March 22nd, 2011. These would have broken the script in any case.
We are currently removing these backups, as they are redundant data, and we have rsync'd backups of the frames and trends. We should now have approximately twice the lookback of full frames.
Since Leo was trying to demo his LIGO Data Listener code, he noticed that there was and NDS2 issue. The NDS2 guy (JZ) noticed that the FrameBuilder had an issue.
We investigated. At 4PM on Dec 31, the GPS timestamp of the frame file names started to be recorded wrong. In fact, it started to give it a file name matching the correct time from 1 year in the past.
So that's our version of the Y2011 bug. Here's the 'ls' of /frames/full:
drwxr-xr-x 2 controls controls 252K Dec 26 03:59 9773
drwxr-xr-x 2 controls controls 260K Dec 27 07:46 9774
drwxr-xr-x 2 controls controls 256K Dec 28 11:33 9775
drwxr-xr-x 2 controls controls 252K Dec 29 15:19 9776
drwxr-xr-x 2 controls controls 244K Dec 30 19:06 9777
drwxr-xr-x 2 controls controls 188K Dec 31 16:00 9778
drwxr-xr-x 2 controls controls 148K Jan 1 08:53 9463
drwxr-xr-x 2 controls controls 260K Jan 2 12:39 9464
drwxr-xr-x 2 controls controls 252K Jan 3 16:26 9465
drwxr-xr-x 2 controls controls 248K Jan 4 20:13 9466
drwxr-xr-x 2 controls controls 36K Jan 5 00:22 9467
controls@fb /frames/full $
The culprit is the directory who's name starts out as 9463 whereas it should be 9779.
Email from Alex:
Turned out to be the lack of current year information in the IRIG-B signal
received by the Symmetricom GPS card in the frame builder machine caused
this. I have added a constant in daqdrc to bring the seconds forward:
controls@fb /opt/rtcds/caltech/c1/target/fb $ grep symm daqdrc
Hopefully we will be upgrading to the newer timing system at the 40M this
year, so this will not happen again next year.
Doing an 'ls -lrt' in /frames/full/ now shows that the names are correct:
drwxr-xr-x 2 controls controls 249856 Dec 30 19:06 9777
drwxr-xr-x 2 controls controls 192512 Dec 31 16:00 9778
drwxr-xr-x 2 controls controls 151552 Jan 1 08:53 9463
drwxr-xr-x 2 controls controls 266240 Jan 2 12:39 9464
drwxr-xr-x 2 controls controls 258048 Jan 3 16:26 9465
drwxr-xr-x 2 controls controls 253952 Jan 4 20:13 9466
drwxr-xr-x 2 controls controls 151552 Jan 5 13:54 9467
drwxr-xr-x 2 controls controls 12288 Jan 5 15:57 9783
Just a proof that the DAQ is working - ran DTT on nodus from 3 hours ago.
We looked into the /frames situation a bit tonight. Here is a summary:
Plan of action:
BTW - the last chiara (shared drive) backup was October 16 6 am. dmesg showed a bunch of errors, Koji is now running fsck in a tmux session on chiara, let's see if that repairs the errors. We missed the opportunity to swap in the 4TB backup disk, so we will do this at the next opportunity.
DTT stopped working for recent data. An 'ls' in the frames/full/ directory reveals:
drwxr-xr-x 2 controls controls 258048 Feb 3 12:26 9807
drwxr-xr-x 2 controls controls 258048 Feb 4 16:13 9808
drwxr-xr-x 2 controls controls 262144 Feb 5 19:59 9809
drwxr-xr-x 2 controls controls 258048 Feb 6 23:46 9810
drwxr-xr-x 2 controls controls 258048 Feb 8 03:33 9811
drwxr-xr-x 2 controls controls 262144 Feb 9 07:19 9812
drwxr-xr-x 2 controls controls 253952 Feb 10 11:06 9813
drwxr-xr-x 2 controls controls 266240 Feb 11 14:53 9814
drwxr-xr-x 2 controls controls 266240 Feb 12 18:39 9815
drwxr-xr-x 2 controls controls 266240 Feb 13 22:26 9816
drwxr-xr-x 2 controls controls 262144 Feb 15 02:13 9817
drwxr-xr-x 2 controls controls 253952 Feb 16 05:59 9818
drwxr-xr-x 2 controls controls 241664 Feb 17 09:46 9819
drwxr-xr-x 2 controls controls 28672 Feb 17 12:22 9820
drwxr-xr-x 2 controls controls 32768 Feb 17 15:06 6663
drwxr-xr-x 2 controls controls 73728 Feb 17 23:39 6664
controls@fb /frames/full $ date
Thu Feb 17 23:39:27 PST 2011
There are at least 5 free DAC channels (4 if you discount the one channel from these that I am hijacking) available in the 1Y2 electronics rack.
Jamie's nice wiring diagram shows the topology - the actual DAC card sits in 1Y3 inside the c1lsc expansion chassis (while the c1lsc frontend itself is in 1X4). The output of the DAC goes via SCSI to an interface box (D080303) and then to some dewhitening/AI boards (D000316). There are a total of 16 DAC channels available, out of which 8 are used for the TTs, 2 are used for the DAFI model, and one is labeleld "From c1ioo 1X2" (I don't know what this one is for). So I'm going to use some of these channels for measuring the coupling of oscillator noise and intensity noise to MICH in the DRMI lock.
The de-whitening/AI board seems to be old - it has 2x 800Hz Butterworth LPFs and no notch for the clock frequency, but maybe this doesn't matter for the tests I have in mind. The AI board available on 1X2 is more modern but routing the DAC channels from 1Y2 to it is going to be some work.
I'm going to add my testpoint to c1daf given that it seems to be the least critical model on c1lsc.
EDIT: testpoints added to c1daf don't show up in the list of available channels - there was some issue with this model while we were getting the new RTCDS going. So I'm moving my temporary testpoint to c1cal instead.
its an acquired taste, but its a must since we're sending an interferometer to India
Free swing of ITMY started at
Tue Sep 6 17:41:43 PDT 2011
I think Kiwamu accidentally restarted this kick at 17:48:02 PDT.
The free swinging spectra of ITMs, ETMs, BS, PRM and SRM were measured last night in order to make sure that nothing wrong have happened by the wiping.
I think there are nothing wrong with ITMs, ETMs, BS, PRM and SRM successfully.
For the comparison, Yoichi's figure in his elog entry of Aug.7 2008 is good, but in his figure somehow PRM spectrum doesn't look correct.
Anyway, compared with his past data, there are no significant changes in the spectra. For PRM which has no counterpart to compare with, its shape of spectra looks similar to any other spectra. So I think PRM is also OK. The measured spectra are attached below.
Motivated by the strange pitch/yaw coupling behavior we ran into while doing diagonalization, we looked at the oplev pitch and yaw free swing spectra for all 4 test masses (see attachment 1). We saw the same behavior there: At the peak frequencies for the angular degress of freedom, the oplevs saw significant contributions from both pitch and yaw. We also examined the phase between pitch and yaw at these peaks and found that consistently, pitch and yaw were in phase at one of the resonance frequencies and out of phase at the other (ignoring the pos and side peaks).
This corresponds physically to angular motion about some axis that is diagonal, ie not perfectly vertical or horizontal. If we trust the oplev calibration, and Eric says that we do, then the angle of this axis of rotation with the horizontal (pitch axis) is
Where Y and P are yaw and pitch ASD values. This will always give an angle between 0 and 90 degrees; which quadrant the axis of rotation occupies can be dermined by looking at the phase between pitch and yaw at the same frequencies. 0 phase means that the axis of rotation lies somewhere less than 90 degrees counterclockwise from the horizontal as viewed from the AR face of the optic, and a phase of 180 degrees means the axis is clockwise from horizontal (see attachment 2). Qualitatively, these features show up the same way for segments of data taken at different times. In order to get some quantitative sense of the error in these angles, we found them using spectrogram values with a bandwidth of 0.02 Hz averaged over 4000 seconds.
Results (all numbers in degrees unless otherwise specified):
peak 1 ( 0.692 Hz):
ptich/yaw phase: -179.181
peak 2 ( 0.736 Hz):
pitch/yaw phase: 0.0123677
peak 1 ( 0.502 Hz):
ptich/yaw phase: -179.471
peak 2 ( 0.688 Hz):
pitch/yaw phase: -0.43991
peak 1 ( 0.73 Hz):
ptich/yaw phase: -0.227034
peak 2 ( 0.85 Hz):
pitch/yaw phase: -179.856
peak 1 ( 0.724 Hz):
ptich/yaw phase: 6.03312
peak 2 ( 0.844 Hz):
pitch/yaw phase: -176.838
ETMY and ITMX both show a more significant (~4x) contribution from pitch on one peak, and from yaw on the other. This is reflected in the fact that they each have one angle somewhat close to 0 (below 30 degrees) and one close to 90 (above 60 degrees). The other two test masses don't follow this rule, meaning that the 2 angular frequency peaks do not correspond to pitch and yaw straightforwardly.
Also, besides ITMX, the axes of rotation are at least several degrees away from being perpendicular to each other.
I have the DRMI free swinging right now, since it's not really locking. Looking at these time series in the attached pdf, particularly around time=1.15, it would be super handy to trigger the SRCL degree of freedom on AS110 after the PRMI is triggered on POP22.
In this night, I checked the free swinging spectra of ETMX to make sure nothing wrong with ETMX by the wiping.
Compared with the past (Aug.6 2008), the spectra of ETMX doesn't show significant change.
Successfully the wiping activity didn't change its configuration so much and didn't bring bad situations.
(bad situation means for example, the suspended components hit some others).
The spectra of ETMX by DTT are attached. Also you can see the past spectra in Yoichi's entry.
Yoichi's data was taken during the air-pressure condition, so it's good for comparing.
Actually I compared those data by my eyes, because I could not get the past raw data somehow.
The resonant frequencies and their typical height changed a little bit, but I think those are not significant.
NOTE: In the figure, pitch and yaw modes (~0.57Hz and ~0.58Hz) look like having a smaller Q-factor than the past.
To take the free swinging Michelson measurements for the interferometer calibration Jamie aligned the beam splitter with ITMX and ITMY. I recorded the GPS time (1027827100 and for several hundred seconds later) when the Michelson was aligned in order to look at the correct data. I then copied the python script nds-test.py from Jamie, and modified it to take and plot data from C1:LSC-AS55_Q_ERR_DQ offline. I used dataviewer to verify that C1:LSC-AS55_Q_ERR_DQ and C1:LSC-AS55_Q_ERR were recording the same signal, and to check that I was taking the correct data with NDS. Taking data online worked as well, but it was easier to use a time when the Michelson was known to be free-swinging and take data offline. Attached is some sample data while free-swinging, with time in GPS time.
I ran "freeswing all" at Fri Aug 19 01:09:28 PDT 2011 (997776583) and "opticshutdown" as well.
I noticed that Chiara's backup HD (which has a capacity of 1.8TB, vs the main drives 2TB) was near to getting full, meaning that we would soon be without a local backup.
I freed up ~200GB of space by compressing the autoburt snapshots from 2012, 2013, 2014. Nothing is deleted, I've just compressed text files into archives, so we can still dig out the data whenever we want.
I just started a freeswing all, as a final check before we pump:
Wed Sep 7 00:43:21 PDT 2011
Wed Sep 7 00:43:32 PDT 2011
WATCHDOGS WILL BE RESET 5 HOURS AFTER THIS TIME
sleeping for 5 hours...
Jamie: Please do a quickie analysis (at least for the ITMs) before helping Steve with the heavy doors.
I closed the PSL shutter.
Both ITM chambers were checked for tools, so there should be nothing left to do but put the heavy doors on, and begin pumping.
I modified the script freeSwing.py to use damping loop output switches to free the optic instead of watchdog or coil output filters. This ensures that the free swing test is being done at the nominal position of the optic. I started tests for LO1, LO2, As2, As4, PR2, PR3, and SR2 in a tmux session names freeSwing on rossa.
Note: LO2 face OSEMs are hardly sensitive to any motion right now due to excessive pitch offset required for LO beam. We should relieve this offset to LO1 and rerun this test later.
This work is now complete. The box was characterized and re-installed in 1X2. I am able to (briefly) lock the IMC and see PDH fringes in POX and POY so the lowest order checks pass.
Even though I did not deliberately change anything in the 29.5 MHz path, and I confirmed that the level at the output is the expected 13 dBm, I had to lower then IN1 gain of the IMC servo by 2dB to have a stable lock - should confirm if this is indeed due to higher optical gain at the IMC error point, or some electrical funkiness. I'm not delving into a detailed loop characterization today - but since my work involved all elements in the RF modulation chain, some detailed characterization of all the locking loops should be done - I will do this in the coming week.
After tweaking the servo gains for the POX/POY loops, I am able to realize the single arm locks as well (though I haven't dont the characterization of the loops yet).
I'm leaving the PSL shutter open, and allowing the IMC autolocker to run. The WFS loops remain disabled for now until I have a chance to check the RF path as well.
Unrelated to this work: Koji's swapping back of the backplane cards seems to have fixed the WFS2 issue - I now see the expected DC readbacks. I didn't check the RF readbacks tonight.
Update 7 Dec 2020 1 pm: A ZHL-2 with heat sink attached and a 11.06 MHz Wenzel source were removed from the box as part of this work (the former was no longer required and the latter wasn't being used at all). They have been stored in the RF electronics cabinet along the east arm.
This turned out to be a much more involved project than I expected. The layout is complete now, but I found several potentially damaged sections of cabling (the stiff cables don't have proper strain relief near the connectors). I will make fresh cables tomorrow before re-installing the unit in the rack. Several changes have been made to the layout so I will post more complete details after characterization and testing.
I was poring over minicircuits datasheets today, and I learned that the minicircuits bandpass filters (SBP10.7 and SBP60) are not bi-directional! The datasheet clearly indicates that the Male SMA connector is the input and the Female SMA connector is the output. Almost all the filters were installed the other way around 😱 . I'll install them the right way around now.
The freq fluctuation of the beat note has been measured with the following condition
- The output of the freq divider is already calibrated to have the unit of MHz.
- The transfer function between the analog PFD channel and the digital PFD output was measured to be -23dB = 0.7.
The gain of the XARM-FINE channel was changed to 0.7 such that the output is calibrated in MHz.
- I have not checked the analog noise level of the analog PFD path. We may need more whitening gain (by icreasing the gain of SR560).
- The analog PFD is always better than the digital PFD above 20Hz.
- Both the digital and analog PFD showed good agreement below 20Hz.
Note the measurement was not simultaneous.
- When the arm is locked with the ETMX being actuated , the fluctuation of the arm length must be stabilized by a huge factor
(~10^5 according to Kiwamu's entry) However, we only could see the stabilization factor of 30.
As this residual is the difference of the freq noise felt by the IR and the green,
this is a real issue to be tackled.
- The RMS fluctuations of the arm with and without the IR beam being locked are 2MHz and 0.1MHz,
which correcponds to the arm length motion of 250nm and 13nm, respectively.
Ed: I had to use 532nm in stead of 1064nm. The correct numbers are 130nm and 7nm.
- Without the IR locked, The typical peak-to-peak fluctuation of the beat freq was 10MHz.
I found that some flakiness of the beat signals comes from the RF components for the beat detection.
They are touching the racks in an indefinite way. If we move the components the output of the analog PFD
Once Kiwamu is back I will ask him to clean up all of the green setting in an appropriate way.
I uploaded all the material about the RF frequency Generation Box into the SVN under the path:
I structured the directory as shown in this tree:
I'm quickly describing in a section of the Rf system upgrade document with LIGO # T1000461.
I completed a LIGO document describing design, construction and characterization of the RF System for the 40m upgrade.
It is available on the SVN under https://nodus.ligo.caltech.edu:30889/svn/trunk/docs/upgrade08/RFsystem/RFsystemDocument/
It can also be found on the 40m wiki (http://lhocds.ligo-wa.caltech.edu:8000/40m/Upgrade_09/RF_System#preview), and DCC under the number T1000461.
I measured the amplitude noise of the source outputs and the EOM outputs of the Frequency Generation box.
the setup I used is shown in this diagram:
(NB It's important that the cables from the splitter to the RF and LO inputs of the mixers are the same length).
The results of the measurements are shown in the following plot:
1) both Crystals (29.5MHz and 11MHz) have the same noise
2) the 55MHz source's noise is bigger than the 11 MHz (~2x): the frequency multiplication and amplification that happen before it, add extra noise
3) the noise at EOM outputs is ~2x bigger than that of the relative sources
When I have the chance, I'll plot the results of my calculations of expected noise and compare them with the measurements.
Here are the results of my phase noise measurements on the 7 outputs of the Frequency Generation Box. (BIN=95L applied by DTT). See attached pdf for a higher definition picture.
The plot shows that the phase noise of the 11 MHz outputs (Source, EOM modulation signal, Demodulation signal) is as low as that of the Marconi. The Marconi is limiting my measurement's resolution.
The mode cleaner signal's oscillator (29.5 MHz output, blue trace) is higher than the 11MHz above 1KHz.
The 55MHz signals have all the same phase noise (traces overlapped), and that is higher than the 11 MHz ones from about 100Hz up. i don't know what's going on.
I need to use the spare 11MHz Wenzel crsytal to have a better reference source for the measurement.
I finished assembling the frequency generation unit for the upgrade. I tested it through to check that the power levels are as expected at the various connection (see attached png, showing in black the design power values, and in red the measured ones).
Because of some modifications made on the design along the construction, I have to recalculate the SNR along the lines.
I can now start to measure phase noise and distortion harmonics.
A document with a description of the design and the results of the characterization measurements will be available in the end.
Attached is the timeline for Frequency Offset Locking related activities. All activities will be done mostly in morning and early afternoon hours.
Unfortunately the order placed for beam samplers last week did not go through. These will be used at the X and Y end tables to dump the unwanted light appropriately. Since they will not be here until Tuesday, I revised the timeline for FOL related activities accordingly.
I finished the polishing in the scripts/FOL directory, this is the current status and this post replaces my two previous posts on the subject:
So, I actually took these measurements last week, but I didn't get around to making nice plots and things until now. I figured the time while I wait for the spectrum analyzer to do its thing was a good time.
Having been unable to locate the SR785 and also unsure how to connect it to a computer speaker (and also unable to find a free one), I downloaded a demo of a function generator onto my computer and just used that. (Same thing I used to do the swept sine that created the frequency power response plots I posted last week.) I set the program to a number of different frequencies and had the other end of the cable hooked into the oscilloscope to see a) if I could pick out the frequency and b) see how the magnitude of the microphone output varied with the frequency.
The first set of measurements I took, I didn't realize that I could increase the output power of the function generator. Because the generated sound at the default setting was relatively quiet, the oscilloscope traces were pretty chaotic, so I usually froze the trace so that I could look at it better. I ended up with a lot of weird jumps in the magnitude, but I later realized that there was a lot of beating going on at some frequencies, and the amplitude changes were probably much more drastic for the -20 dB sounds than the 6 dB sounds, since it was closer in amplitude to the surrounding noises. So, I've included that data set in my plots for the sake of completeness, but I'm pretty sure that it is useless.
Once I realized I could increase the power output for the signal generator, I took a set of data with and without the voltage divider at 6 dB. There was a cluster of frequencies that showed significant beating around 1700-3000 Hz in the data WITH the voltage divider, but I did not see any clear beating in the data WITHOUT. In the plots, I simply plotted up the highest and lowest amplitudes I measured for the frequencies with significant beating, since it was obviously hard to tell what the amplitude would have been without any background noise. In the w/o volt. div. set, although I didn't see any obvious beat patterns, the measured amplitudes did jump slightly at the frequencies that showed beats with the voltage divider. So, perhaps I was just not seeing them, but they influenced my amplitude measurements? I'm not sure if it would be possible for the voltage divider itself to cause beat frequencies.
(Note: the amplitudes measured were from zero to peak, as the oscilloscope I was using wouldn't show a big enough vertical range to easily measure the peak-to-peak voltage difference.)
I've attached two plots of my measurements. One has a regular x-scale and includes all the measurements. The second has a logarithmic x-scale and omits the 20 Hz points. I had some troubles being able to pick out the 20 Hz signal on the oscilloscope... I don't know if my computer speakers just don't work well at that frequency or what, but either way, those points seemed highly suspect, and omitting them from the log plot allowed me to spread things out more.
One thing I'm not sure about is the 3000 Hz point. It was one of the ones with a beat frequency (~130 Hz), and the amplitudes were pretty low. The corresponding point from the non-voltage-divider data set is also low. So, I'm not sure what's happening there.
The one thing that I do think is quite clear is that the 1000 Hz drop-off in power when the microphone is connected to the ADC has nothing to do with the voltage divider. Beat issues aside, the shapes are very similar (pay no attention to the absolute scale... obviously, the voltage responses with and without the voltage divider were very different, and I just scaled them to fit in the same plot).
Update: Jenne pointed out that I was not absolutely clear about the voltage scale in my plots. The GREEN and BLUE points are on a mV scale, and the RED points are on a 10mV scale. I should probably redo the plots in Matlab in eventuality, since Excel is hard to use if you want to do anything that is not extremely basic with your plots, but this was my solution for the time being. So, the fact that the RED points, which are the data taken WITHOUT the voltage divider, are lower than the GREEN ones does not in any way indicate that I measured lower voltages when the voltage divider was not used.
Also, a to do list:
- Many of the beat frequencies I picked out were veeeeery slow, indicating that something is going at a frequency that is very close to the arbitrary frequencies I chose to sample, which is a little strange. That, combined with the fact that I saw clear beats with the voltage divider but not without leads me to believe that it may be worth investigating the frequency response of the voltage divider itself.
- Redo the measurements near the anomalous 3000 Hz point with a higher density of sampled frequencies to try to see what the heck is going on there.