40m QIL Cryo_Lab CTN SUS_Lab TCS_Lab OMC_Lab CRIME_Lab FEA ENG_Labs OptContFac Mariner WBEEShop
  40m Log, Page 152 of 344  Not logged in ELOG logo
ID Date Author Type Category Subjectup
  9762   Sat Mar 29 00:11:39 2014 KojiConfigurationGeneralNTP setting on nodus

FB: /etc/ntp.conf was updated as below such that it refers nodus and also caltech server when nodus is not available

server 192.168.113.200
server 131.215.239.14

ntpd was restarted and

diskless RT machines: they are booted from /diskless/root on fb.
Therefore /diskless/root/etc/ntp.conf was updated in the same way as above.
When the machines are rebooted, this setting will be active.

control machines:

now they are referring nodus and other external servers too.

  2005   Fri Sep 25 19:56:08 2009 ranaConfigurationComputersNTPD restarted on c1dcuepics (to fix the MEDM screen times)
restarted ntp on op440m using this syntax
>su
>/etc/init.d/xntpd start -c /etc/inet/ntp.conf

gettting the time on scipe25 (for the MEDM screen time) working was tougher. The /etc/ntp.conf file was pointing
to the wrong server. Our NAT / Firewall settings require some of our internal machines to go through the gateway
to get NTPD to work. Curiously, some of the linux workstations don't have this issue.
The internal network machines should all have the same file as scipe25's /etc/ntp.conf:

server nodus

and here's how to check that its working:

[root@c1dcuepics sbin]# ./ntpq -p
     remote           refid      st t when poll reach   delay   offset  jitter
==============================================================================
 nodus.ligo.calt 0.0.0.0         16 u    -   64    0    0.000    0.000 4000.00
*nodus.ligo.calt usno.pa-x.dec.c  2 u   29   64  377    1.688  -65.616   6.647
-lime7.adamantsy clock.trit.net   3 u   32   64  377   37.448  -72.104   4.641
-montpelier.ilan .USNO.           1 u   19   64  377   18.122  -74.984   8.305
+spamd-0.gac.edu nss.nts.umn.edu  3 u   28   64  377   72.086  -66.787   0.540
-mighty.poclabs. time.nist.gov    2 u   30   64  377   71.202  -61.127   4.067
+monitor.xenscal clock.sjc.he.ne  2 u   16   64  377   11.855  -67.105   6.368
  4162   Mon Jan 17 04:10:10 2011 ranaConfigurationComputersNTPD restarted on c1dcuepics (to fix the MEDM screen times)

I installed NTPD on Megatron (its Ubuntu, so different from the CentOS workstations). Here's the terminal cap to show that its actually working:

megatron:/etc>sudo /etc/init.d/ntp restart
 * Stopping NTP server ntpd                                              [ OK ]
 * Starting NTP server ntpd                                              [ OK ]
megatron:/etc>/etc/init.d/ntp status
 * NTP server is running
megatron:/etc>ntpq -p
     remote           refid      st t when poll reach   delay   offset  jitter
==============================================================================
 nodus.martian   204.123.2.72     2 u    7   64    1    0.217    3.833   0.001
 europium.canoni 193.79.237.14    2 u    6   64    1  155.354    3.241   0.001
megatron:/etc>date
Mon Jan 17 04:07:08 PST 2011

Along the way, I also updated the /etc/inet/ntp.conf file for nodus. It was using the USNO as a NTP server and I've pointed it to the Caltech NTP server as per the IMSS NTP page.

  4283   Mon Feb 14 01:40:14 2011 KojiOmnistructureCDSName of the green related channels

I propose to use C1:ALS-xxx_xxx for the names of the green related channels, instead of GCV, GCX, GCY, GFD...

Like C1:SUS or C1:LSC, we name the channels by the subsystems first, then probably we can specify the place.

We can keep the names of the processes as they are now.

  965   Thu Sep 18 14:36:54 2008 josephbConfigurationComputersName server and Epics
The problems Rob was experiencing last night was due to part of the setup (or rather testing of the setup) of the new nameserver running on linux1.

The name server was setup on linux1 by doing the following:

1) Installed xorg-x11-xauth via yum which was necessary to get remote x windows to work in linux1

2) Installed xorg-x11-fonts-Type1 in order to get the gui system-config-* programs to work

3) Ran system-config-bind, which created a default set of nameserver files. I unfortunately didn't understand the gui all that well, so I manually edited and added files to these base ones. The base files were generated in /var/named/chroot/etc/ and /var/named/chroot/var/named.

4) I added martian.zone and 113.215.131.in-addr.arpa.zone, named.conf.local, and edited named.conf so it loaded named.conf.local. The martian.zone file acts a forward look up (i.e. give it a name and it returns an IP number like 131.215.113.20). The 113.215.131.in-addr.arpa.zone acts as a reverse look up (i.e. give it an IP number like 131.215.113.20 and it tells you the name). The file named.conf.local merely points to these two files.

Note: One can add or change IP lookup by simply updating these two files. The format should be obvious from the files.

5) I specifically ssh'd in as root to linux1 (using su wasn't sufficient) and then typed "service named start" (without quotes). You can also use "restart" or "stop" instead of "start". This started the name server, giving an [Ok] message.

6) I edited the /etc/resolve.conf file on linux1 so that it pointed to itself first ("nameserver 127.0.0.1" at the top of the file). I also added the line "search martian", which allows one to simply use linux1 as opposed to linux1.martian.

I also edited the /etc/resolve/conf file on linux2, and it seems to resolve names fine.

7) And here is where I broke things. As a test, I moved /etc/hosts to /etc/hosts.bak, and then tested to see if names were being resolved correctly. By using the command host, I determined they were in fact working. I also tested with ssh.

However, something basic didn't like me moving the hosts file. Apparently when a front-end machine needed to reboot, it wouldn't come back up, without any ability to SSH or telnet into them.

With Yoichi and I did quite a bit of debugging this morning and determined the nameserver itself isn't conflicting, merely the lack of the host file was the source of the problem. One theory is that services don't know to go to DNS to resolve host names. I think by modifying the /etc/nsswitch.conf file to include dns as an option for services and other programs, it might work without the host file, however, I'm going to leave that to tomorrow morning which is less likely to interfere with current operations.

As it stands, things are working with the nameserver running and the host file in place.
  204   Wed Dec 19 20:28:27 2007 AndreyDAQPEMNames for all 6 accelerometers have been changed

I eventually changed the names for all 6 accelerometers (see my ELOG entry # 172 from Dec. 05 about my intent to do that).

I removed the word "BS" from their names,
and I changed the word combination "ACC_BS_EAST" in the old name for "ACC_ITMX" in the new name;
as well "ACC_BS_WEST" is now replaced by "ACC_ETMX".
(the reasoning behind such a change should become clear from my ELOG entry #172).

New accelerometer names are:
(note: there are no spaces (nowhere!) in the names of accelerometers, but ELOG replaces ": P" written without a space by a strange symbol Tongue)

C1 : PEM - ACC _ ETMX _ X ;
C1 : PEM - ACC _ ETMX _ Y ;
C1 : PEM - ACC _ ETMX _ Z ;
C1 : PEM - ACC _ ITMX _ X ;
C1 : PEM - ACC _ ITMX _ Y ;
C1 : PEM - ACC _ ITMX _ Z .

One can find them in "C1 : PEM - ACC" in Dataviewer.

  6980   Tue Jul 17 11:28:29 2012 JenneUpdateASSNames of DoF filters in ASS wrong

The names of the DoF filters in the ASS loop were wrong.  The filters themselves were correct (low pass filters at super low freq, for the Lock-in), but the names were backward.

Our convention is to name filters "poles:zeros", but they had been "z:p".  The names of FM1 in all the DoF filter banks are now fixed.

  973   Fri Sep 19 11:21:45 2008 josephbConfigurationComputersNameserver and Rosalba
I tried modifying the nsswitch.conf file to include going to dns in addition to local files for everything (services, network, etc) and then moving the /etc/hosts file to /etc/hosts.bak. Unfortunately, this still didn't allow front-ends to reboot properly. So I'm not sure what is using the hosts file, but whatever it is, is apparently important. After the test I placed the hosts file back and reverted the nsswitch.conf file.

I also noticed that Rosalba was having problems connecting to the network. This apparently was because I had shut down the dhcp server on the NAT router, as had been discussed at the meeting on Wednesday.

To fix this, I modified the /etc/sysconfig/network-scripts/ifcfg-eth1 file to fix rosalba's ip as 131.215.113.24 (which doesn't seem to be in use). I also updated rosalba's /etc/resolv.conf file to point at linux1's name server, and two additional name servers as well, and added the "search martian" line. I modified the /etc/sysconfig/network-scripts/ifcfg-eth0 file so the built in network card doesn't come up automatically, since its currently not plugged into anything. Lastly, I added rosalba and its IP to linux1's name server files.
  8479   Tue Apr 23 22:10:54 2013 ranaUpdateComputersNancy

controls@rosalba:/users/rana/docs 0$ svn resolve --accept working nancy
Resolved conflicted state of 'nancy'

  15745   Wed Dec 23 10:13:08 2020 gautamUpdateCDSNear term upgrades

Summary:

  1. There appears to be insufficient number of PCIe slots in the new Supermicro servers that were bought for the BHD upgrade.
  2. Modulo a "riser card", we have all the parts in hand to put one of the end machines on the Dolphin network. If the Rogue Master doesn't improve the situation, we should consider installing a Dolphin card in the c1iscex server and connecting it to the Dolphin network at the next opportunity. 

Details:

Last night, I briefly spoke with Koji about the CDS upgrade plan. This is a follow up.

Each server needs a minimum of two peripheral devices added to the PCIe bus:

  • A PCIe interface card that connects the server to the Expansion Chassis (copper or optical fiber depending on distance between the two).
  • A Dolphin or RFM card that makes the IPC interconnects. 
  • I'm pretty certain the expansion chasiss cannot support the Dolphin / RFM card (it's only meant to be for ADCs/DACs/BIO cards). At least, all the existing servers in the 40m have at least 2 PCIe cards installed, and I think we have enough to worry about without trying to engineer a hack.
  • I attach some photos of new and old Supermicro servers to illustrate my point, see Attachment #1

As for the second issue, the main question is if we had an open PCIe slot on the c1iscex machine to install a Dolphin card. Looks like the 2 standard slots are taken (see Attachment #1), but a "low profile" slot is available. I can't find what the exact models of the Supermicro servers installed back in 2010 are, but maybe it's this one? It's a good match visually anyways. The manual says a "riser card" is required. I don't know if such a riser is already installed. 

Questions I have, Rolf is probably the best person to answer:

  1. Can we even use the specified host adaptor, HIB25-X4, which is "PCIe Gen2", with the "PCIe Gen3" slots on the new Supermicro servers we bought? Anyway, the fact that the new servers have only 1 PCIe slot probably makes this a moot point.
  2. Can we overcome this slot limitation by installing the Dolphin / RFM card in the expansion chassis?
  3. In the short run (i.e. if it is much faster than the full CDS shipment we are going to receive), can we get (from the CDS test stand in Downs or the site) 
    • A riser card so that we may install the Gen 1 Dolphin card (which we have in hand) in the c1iscex server?
    • A compatible (not sure what PCIe generation we need) PCIe host to ECA kit so we can test out the replacement for the Sun Microsystems c1ioo server?
    • A spare RFM card (VMIC 5565, also for the above purpose). 
  4. What sort of a test should we run to test the new Dolphin connection? Make a "null channel" differencing the same signal (e.g. TRX) sent via RFM and Dolphin? Or is there some better checksum diagnostic?
Attachment 1: IMG_0020.pdf
IMG_0020.pdf
  14389   Tue Jan 8 10:27:27 2019 gautamUpdateGeneralNear-term in-chamber work

Here is a list of tasks I think we should prioritize for the next two weeks. The idea is to get back to the previous state of being able to do single arm, PRMI-on-carrier and DRMI locking, before making further changes.

Once the new folding mirrors arrive, I'd like to modify the SRC length to allow locking in the signal-recycled config as opposed to RSE. Still need to do the detailed layout, but I think the in-vacuum layout will work. In that case, I'd like to also move the OMC and OMMT to the IY table, and also move the in-air AS photodiodes to the IY in-air optical table. This is why I've omitted the OMC alignment from this near-term list, but if we want to not move the OMC, then we probably should add alignment of the AS beam to the OMC to this list.

List of in-chamber tasks for 1/2019

Chamber Task(s)
EY
  • Clean ETMY optic and suspension
  • Put ETMY suspension back in place, recover Y-arm cavity alignment
  • Remove any residual hardware from unused heater setup
  • Restore parabolic heater setup, center radiation pattern as best as possible on ETMY
  • Check beam position on IPANG steering mirror
IY
  • Clean ITMY optic and suspension cage
  • Restore ITMY suspension, recover Y arm cavity alignment.
  • Check position of AS beam on OM1/OM2
BS/PRM (if we decide to open it)
  • Replace BS/PRM Oplev HeNe, bring the beam in and out of vacuum with beam well centerd on in-vacuum mirrors (can take this opportunity to fix the in-air layout as well to minimize un-necessary steering mirrors)
  • Check position of AS beam on OM3/OM4, adjust if necessary
  • Check position of IPPOS and IPANG beams on their respective steering optics
OMC (if we decide to open it)
  • Check position of AS beam on OM5/OM6
  • Ensure AS beam exits the vacuum cleanly
  7110   Tue Aug 7 23:30:58 2012 ranaUpdateEnvironmentNearby EQ

 Just felt an EQ. Impulse moved some vertical blinds by several mm.

Tue Aug 07 23:26:06 2012 

  7111   Tue Aug 7 23:33:34 2012 DenUpdateEnvironmentNearby EQ

Quote:

 Just felt an EQ. Impulse moved some vertical blinds by several mm.

Tue Aug 07 23:26:06 2012 

 All optics except MC2 and ETMX are crazy

watchdogs.png      mc1.png    gur.png

  7387   Fri Sep 14 12:51:43 2012 JenneUpdateSUSNeed risers for active TTs

I was helping Den get started in the cleanroom yesterday, and I noticed that the new active TTs, like the old passive ones, are set to be 4" from the table.  So, like the old ones, we need 1.5" risers to get the center of the mirror up to our in-vac 5.5" beam height.  I didn't see any risers in there when I was looking around. 

Steve says he still has the drawing that he gave to the shop for the old tip tilts, so he'll double check that the dimensions are the same, and then ask the shop to make 4 more.

  8781   Fri Jun 28 02:23:00 2013 JenneUpdateLSCNeed to measure sensing matrix at REFL165

[Lisa, Rana, Jenne]

Lisa asked to see a model of the PRMI sensing matrix with REFL165 included, in the hopes that it wouldn't be as degenerate as REFL33.

SensMatModel_28June2013_InclREFL165.png

The conclusion, immediately after looking at this, is that I should make sure the REFL beam is nicely aligned onto the REFL165 PD (Koji did some tests, swapping out the REFL165 resonant PD with a broadband PD, and I don't remember if he aligned beam back onto the REFL165 PD).  Then, I need to measure the PRMI sensing matrix, including REFL165.  Hopefully, it is similar to the model, and we can use it as our 3f diode for locking.

  8783   Fri Jun 28 12:15:09 2013 KojiUpdateLSCNeed to measure sensing matrix at REFL165

There is no sensible REFL165 PD in the lab. I am supposed to prepare a new version of REFL165 using prototype BBPD.

  2932   Fri May 14 12:14:26 2010 josephbUpdateCDSNeed to track down old code for lsc system and remove them

I'm currently in the process of tracking down what legacy code is interfering with the new lsc model.

It turns out if you change the name of lsc file to something else (say scx as a quick test for example), it runs fine.  In fact, the lsc and scx GDS_TP screens work in that case (since they're looking at the same channels).  As one would expect, running them both at the same time causes problems.  Note to self, make sure the other one is killed first.  It does mean the lsc code gets loaded part way, but doesn't seem to communicate on EPICs or to the other models.  However, I don't know what existing code is interfering.  Currently going trhough the target directories and so forth.

  1993   Fri Sep 18 16:26:02 2009 steveSummaryPSLNeslab chiller is OK

Rob found puddles of water very close to the chiller during lunch time. We raised the unit and took the side cover off. All surfaces were dry and the water level in the tub normal.

Later on we discovered that one of the Vons distilled water bottle was leaking. Jenne and I checked for excess amount of condensing water droplets inside the MOPA box.

On the bare,not insulated tubing and valve are loaded with droplets of water. Relative humidity is 44% at 24 C and HEPA filter speed set to 80 V in the enclosure.

 

  3526   Mon Sep 6 10:08:10 2010 AlbertoConfigurationComputersNetgear Network Switch fan broken.

The Netgear Network Switch in the top shelf of Nodus' rack has a broken fan. It is the one interfaced to the Martian network.

The fan must have broken and it is has now started to produce a loud noise. It's like a truck was parked in the room with the engine running.

Also the other network switch, just below the Netgear, has one of its two fans broken. It is the one interfaced with the General Computer Side.

I tried to knock them to make the noise stop, but nothing happened.

We should consider trying to fix them. Although that would mean disconnecting all the computers.

  3595   Wed Sep 22 22:22:12 2010 KojiConfigurationComputersNetgear Network Switch fan broken.

Net switch mumbo-jumbo:

Although Rana is going to buy a replacement for the Netgear Switch for martian, I opened the lid of the Netgear as the fan already have stopped working.
Also the lid of the other network switch for GC (Black one) was opened as it has a broken fan and a noisy half-broken fan.

I have asked Steve to buy replacement fans. These would also be the replacement of the replacement.

During the work, it seemed that I accidentally toggled the power supply of linux1. It lead lengthy fsck of the storage.
This is why all of the machines which rely on linux1 got freezed. linux1 is back and the machines looked happy now.

If you find any machine disconnected from the network, please consult with me.

Quote:

The Netgear Network Switch in the top shelf of Nodus' rack has a broken fan. It is the one interfaced to the Martian network.

The fan must have broken and it is has now started to produce a loud noise. It's like a truck was parked in the room with the engine running.

Also the other network switch, just below the Netgear, has one of its two fans broken. It is the one interfaced with the General Computer Side.

I tried to knock them to make the noise stop, but nothing happened.

We should consider trying to fix them. Although that would mean disconnecting all the computers.

 

  10117   Tue Jul 1 18:06:13 2014 nichinUpdateComputer Scripts / ProgramsNetwork Analyzer NWAG4395A data acquisition

EDIT: The script and template file have been moved to /opt/rtcds/caltech/c1/scripts/general/netgpibdata/

 

___________________________________________________________________________________________________________

The NEW and IMPROVED script for remotely getting data from Agilent 4395A network analyzer is located at /users/nichin

This script is quite different from the one in Elog 10108 and fetches us both magnitude and phase. There is an added feature of setting the IF Bandwidth.

The network analyzer is located at crocetta.martian (192.168.113.108)

How to run the script:

> python NWAG4395A_data_acq.py [filename.yml]

  1. The script accepts sweep parameters and output options via a .yml file that is written following a template that can be found at /users/nichin/NWAG4395A_parameters.yml
  2. The data obtained is stored as a .dat file and the corresponding details regarding the acquired data is in a .par parameter file
  3. Separate .dat and .par files are created for phase and magnitude of voltage data.
  4. You can choose to get a plot of the data obtained by specifying it in the .yml file. The plots are automatically stored as PDF.
  5. Plots, data and parameter files are all stored in a new folder that is created with a timestamp in its name.
  6. NOTE: Plotting options are only available in computers running numpy versions of 1.6.0 or above.(Currently only Ottavia and Chiara)

Test Run:

I connected a simple 2MHz Low pass filter between the modulation output and signal input of the NA and ran a scan from 100KHz to 20MHz. The script was run from Ottavia.

The expected plot was obtained and is attached here.

Current work:

Setting up the RF switch in rack 1Y1 to select between required PDs and scripts to tell it which channel to choose over the Ethernet.

 

Attachment 1: LPF2_01-07-2014_175443.pdf
LPF2_01-07-2014_175443.pdf
  153   Sun Dec 2 17:37:33 2007 ranaOmnistructureComputersNetwork Cabling in the Office
We all know that we've spent many integrated man hours trying to figure out why our network connections
in the office area don't work. Usually its because of the bad hub around the Tobin/Osamu desk.

I pried open some of the wall conduit today and it looks pretty easy to fish cables through. I think
its time we finally did that. It may be a little disruptive, but I propose we get Larry to come over
and figure out what needs to happen for us to get regular 100 Mbit ports on the walls. These can
then all go over and get connected to a switch in the rack that holds linux1.

Opinions / comments ?
  977   Mon Sep 22 16:51:27 2008 YoichiHowToComputersNetwork GPIB
I was able to make the wireless connected GPIB interface work with SR785.
Now you can download data from SR785 through network, wherever it is located.
Say good bye to floppy disks.

I wrote an installation note in the wiki.
http://lhocds.ligo-wa.caltech.edu:8000/40m/GPIB

I wrote a new script called "netgpibdata.py" which works similarly as "getgpibdata.py".
It is in the 40m svn. Instructions on how to use it is on the above mentioned wiki page.
  983   Tue Sep 23 00:47:24 2008 YoichiHowToComputersNetwork GPIB

Quote:

I wrote a new script called "netgpibdata.py" which works similarly as "getgpibdata.py".
It is in the 40m svn. Instructions on how to use it is on the above mentioned wiki page.


netgpibdata.py is now installed on the controls machines (/cvs/cds/caltech/scripts/general/netgpibdata/netgpibdata.py).
You can use it like,
netgpibdata.py -i 131.215.113.106 -d AG4395A -a 10 -f spectrum01

In this example, data from Agilent 4395A analyzer at GPIB address 10 connected to the GPIB-LAN box with the IP address 131.215.113.106
is downloaded and saved to spectrum01.dat. The measurement parameters are saved to spectrum01.par.
  10984   Fri Feb 6 15:29:29 2015 ericqUpdateCDSNetwork Shenanigans

Just now we had another EPICS freeze. The network was still up; i.e. I could ssh to chiara and fb, who both seemed to be working fine.

I could ping c1lsc successfully, but ssh just hung. fb's dmesg had some daqd segfault messages, so I telnet'ed to daqd and shut it down. Soon after, EPICS came back, but this is not neccesarily because of the daqd restart...

  11867   Wed Dec 9 18:45:58 2015 KojiSummaryGeneralNetwork Topology Check

[Eric Q, Gautam, Koji]

We went through the network connections to produce the mapping of the instruments.
Gautam summarized the notes into a spread sheet. See attachments.

We didn't find any irregular connections except for the connection of NETMGR port of c1ioo to Martian Network.
This cable was removed.

Attachment 1: Network_topology_9Dec2015.xlsx
Attachment 2: Network_topology_9Dec2015.pdf
Network_topology_9Dec2015.pdf Network_topology_9Dec2015.pdf Network_topology_9Dec2015.pdf
  12963   Wed May 3 16:00:00 2017 gautamSummaryGeneralNetwork Topology Check

[johannes, gautam]

I forgot we had done this last year already, but we updated the control room network switch labels and double checked all the connections. Here is the status of the connections and labels as of today:

There are a few minor changes w.r.t. labeling and port numbers compared to the Dec 2015 entry. But it looks like there was no IP clash between Rossa and anything (which was one of the motivations behind embarking on this cleanup). We confirmed by detatching the cable at the PC end of Rossa, and noticed the break in the ping signals. Plugging the cable back in returned the pings. Because Rossa is currently un-bootable, I couldn't check the MAC address.

We also confirmed all of this by using the web browser interface for the switch (IP = 192.168.113.249).

Attachment 1: Network_topology_3May2017.pdf
Network_topology_3May2017.pdf
  5114   Thu Aug 4 00:04:52 2011 JennyUpdatePSLNetwork analyzer and PD set up to measure amplitude response of PZT

Today I placed the PDA255 photodiode on the PSL table to catch the small amount of beam power reflected by the second polarizing beam splitter in my setup. I plugged the PD output to the oscilloscope to measure the voltage output and positioned the PD such that the voltage output was maximized. At best I was able to achieve a 300 mV DC output voltage from the PD, (which seems a bit low, as the PD is specified to go from 0 to 5 V and the specifications say that the response becomes nonlinear after 10 mW/cm^2 and my beam has an intensity of approximately 5 mw/cm^2. I would therefore expect to get more beam power but after over an hour of maneuvering, 300 mV was the highest voltage output I could get).

I am planning, tomorrow afternoon, to take a measurement of the amplitude response of the PZT driving the NPRO laser. I moved the 4395 spectrum/network analyzer to near the PSL table and connected the RF output to an RF splitter. I fed one output of that into the PZT and the other output into the R port on the network analyzer. I fed the PD output into the A port. I plan to measure A/R as a function of driving frequency, sweeping from 10 Hz to 30 mHz.

I also worked to improve the mode matching of the NPRO beam coming from the AP table to the reference cavity. I drove the temperature of the NPRO at 0.100 Hz with an amplitude of 0.300 V, which Koji told me corresponds to a 1GHz change in the laser frequency. The transmission from the cavity is being monitored by a camera connected to a TV monitor, and also by a PD connected to an oscilloscope. I then repositioned the second lens in my mode matching setup in an attempt to increase the transmission peaks from the zeroth order spacial mode and decrease the transmission peaks from higher order modes. I may have improved the mode matching slightly but I was unable to improve it significantly.

  5126   Fri Aug 5 18:29:35 2011 JennyUpdatePSLNetwork analyzer and PD set up to measure amplitude response of PZT

Quote:

Today I placed the PDA255 photodiode on the PSL table to catch the small amount of beam power reflected by the second polarizing beam splitter in my setup. I plugged the PD output to the oscilloscope to measure the voltage output and positioned the PD such that the voltage output was maximized. At best I was able to achieve a 300 mV DC output voltage from the PD, (which seems a bit low, as the PD is specified to go from 0 to 5 V and the specifications say that the response becomes nonlinear after 10 mW/cm^2 and my beam has an intensity of approximately 5 mw/cm^2. I would therefore expect to get more beam power but after over an hour of maneuvering, 300 mV was the highest voltage output I could get).

I am planning, tomorrow afternoon, to take a measurement of the amplitude response of the PZT driving the NPRO laser. I moved the 4395 spectrum/network analyzer to near the PSL table and connected the RF output to an RF splitter. I fed one output of that into the PZT and the other output into the R port on the network analyzer. I fed the PD output into the A port. I plan to measure A/R as a function of driving frequency, sweeping from 10 Hz to 30 mHz.

I also worked to improve the mode matching of the NPRO beam coming from the AP table to the reference cavity. I drove the temperature of the NPRO at 0.100 Hz with an amplitude of 0.300 V, which Koji told me corresponds to a 1GHz change in the laser frequency. The transmission from the cavity is being monitored by a camera connected to a TV monitor, and also by a PD connected to an oscilloscope. I then repositioned the second lens in my mode matching setup in an attempt to increase the transmission peaks from the zeroth order spacial mode and decrease the transmission peaks from higher order modes. I may have improved the mode matching slightly but I was unable to improve it significantly.

The ABSL beam had been blocked so that it wouldn't enter the interferometer. I moved the block so that the beam I've been using is unblocked by the beam going to the interferometer is still blocked.

I positioned a fast lens (f=28.7mm) a little over an inch in front of the PDA255 in order to decrease the spot size incident on the PD. I adjusted the rotation angle of the half wave plate to maximize the transmitted power through the PBS to the cavity and minimize the power reflected to my PD. I then adjusted the lens potion to fix the beam on the PD. The voltage output of the PD is now 150mW, but I have the ability to increase the incident power by rotating the wave plate slightly.

Now all I need is to set up the network analyzer again to record the amplitude response to modulating the PZT from 10 Hz to 30 MHz, reduce the input voltage into the analyzer using a DC block.

  5144   Mon Aug 8 20:23:14 2011 JennyUpdatePSLNetwork analyzer and PD set up to measure amplitude response of PZT

Quote:

Quote:

Today I placed the PDA255 photodiode on the PSL table to catch the small amount of beam power reflected by the second polarizing beam splitter in my setup. I plugged the PD output to the oscilloscope to measure the voltage output and positioned the PD such that the voltage output was maximized. At best I was able to achieve a 300 mV DC output voltage from the PD, (which seems a bit low, as the PD is specified to go from 0 to 5 V and the specifications say that the response becomes nonlinear after 10 mW/cm^2 and my beam has an intensity of approximately 5 mw/cm^2. I would therefore expect to get more beam power but after over an hour of maneuvering, 300 mV was the highest voltage output I could get).

I am planning, tomorrow afternoon, to take a measurement of the amplitude response of the PZT driving the NPRO laser. I moved the 4395 spectrum/network analyzer to near the PSL table and connected the RF output to an RF splitter. I fed one output of that into the PZT and the other output into the R port on the network analyzer. I fed the PD output into the A port. I plan to measure A/R as a function of driving frequency, sweeping from 10 Hz to 30 mHz.

I also worked to improve the mode matching of the NPRO beam coming from the AP table to the reference cavity. I drove the temperature of the NPRO at 0.100 Hz with an amplitude of 0.300 V, which Koji told me corresponds to a 1GHz change in the laser frequency. The transmission from the cavity is being monitored by a camera connected to a TV monitor, and also by a PD connected to an oscilloscope. I then repositioned the second lens in my mode matching setup in an attempt to increase the transmission peaks from the zeroth order spacial mode and decrease the transmission peaks from higher order modes. I may have improved the mode matching slightly but I was unable to improve it significantly.

The ABSL beam had been blocked so that it wouldn't enter the interferometer. I moved the block so that the beam I've been using is unblocked by the beam going to the interferometer is still blocked.

I positioned a fast lens (f=28.7mm) a little over an inch in front of the PDA255 in order to decrease the spot size incident on the PD. I adjusted the rotation angle of the half wave plate to maximize the transmitted power through the PBS to the cavity and minimize the power reflected to my PD. I then adjusted the lens potion to fix the beam on the PD. The voltage output of the PD is now 150mW, but I have the ability to increase the incident power by rotating the wave plate slightly.

Now all I need is to set up the network analyzer again to record the amplitude response to modulating the PZT from 10 Hz to 30 MHz, reduce the input voltage into the analyzer using a DC block.

 I rolled the network analyzer over to the PSL table (on the south side). I'm borrowing the DC block from Kiwamu's green locking setup. I'm going to first measure the amplitude response of a low pass filter to made sure that the analyzer is outputting what I expect. Then I will measure the laser PZT amplitude response. I plan to finish the measurement and return the network analyzer to it's usual location tonight.

  16656   Thu Feb 10 14:39:31 2022 KojiSummaryComputersNetwork security issue resolved

[Mike P / Koji / Tega / Anchal]

IMSS/LIGO IT notified us that "ILOM ports" of one of our hosts on the "114" network are open. We tried to shut down obvious machines but could not identify the host in question. So we decided to do a bit more systematic search of the host.

[@Network Rack]
- First of all, we disconnected the optical cables coming to the GC router while the ping is running on the AIRLIGO connected laptop (i.e. outside of the 40m network). This made the ping stopped. This means that the issue was definitely in the 40m.
- Secondly, we started to disconnect (and reconnect) the ethernet cables from the GC router one by one. We found that the ping response stops when the cable named "NODUS" was disconnected.

[@40m IFO lab]
- So we tracked the cable down in the 40m lab. After a while, we identified that the cable was really connected to nodus.

- Nodus was supposed to have one network connection to the martian network since the introduction of the bidirectional NAT router (rather than the old configuration with a single direction NAT router).

- In fact, the cable was connected to "non-networking" port of nodus. (Attachment 1). I guess the cable was connected like this long time, but somehow the ILOM (IPMI) port was activated along with the recent power cycling.

- The cable was disconnected at nodus too. (Attachment 2) And a tape was attached to the port so that we don't connect anything to the port anymore.

Attachment 1: PXL_20220210_220816955.jpg
PXL_20220210_220816955.jpg
Attachment 2: PXL_20220210_220827167.jpg
PXL_20220210_220827167.jpg
  449   Fri Apr 25 13:53:11 2008 josephbSummaryComputersNetwork setup
This is the promised more in detail summary from Andrey's log ID 444.

What we did was go around to each hub, one at a time, unplug the network connection, and figure out which light on which hub went out. We then, went back to the control room, confirmed that we were still able to talk to the devices connected to the hub, and if not, rebooted them. This process was repeated for each hub.

As it stands, the hubs located at the ends of arms (in racks 1X4 and 1Y9) are connected to the really old 24 port 10 Base T hub located in 1Y7. In addition, the 5 port SMC hub is plugged into the 8 port SMC switch in 1Y5 (which actually has enough ports to simply move all the connections over to it, so I'm not sure why there are two...).

All other hubs/switches are connected back to the control room 24 port switch.

Attached is a simple diagram of the network connections for the 40m lab.
Attachment 1: 40m_network_90.pdf
40m_network_90.pdf
  464   Mon May 5 11:04:30 2008 robOmnistructureComputersNetwork setup

Mafalda was not connected to the network, and so our DMF-based seisBLRMS has not been running for ~1 week. I traced this to a broken ethernet cable connecting mafalda to the network switch in the rack next to the B&W printer. This cable has a broken connector at the switch side, which means it can't stay connected if there's any tension. It needs to be replaced.
  13037   Sun Jun 4 14:19:33 2017 ranaFrogsComputersNetwork slowdown: Martians are behind a waterwall

A few weeks ago we did some internet speed tests and found a dramatic difference between our general network and our internal Martian network in terms of access speed to the outside world.

As you can see, the speed from nodus is consistent with a Gigabit connection. But the speeds from any machine on the inside is ~100x slower. We need to take a look at our router / NAT setup to see if its an old hardware problem or just something in the software firewall. By comparison, my home internet download speed test is ~48 Mbit/s; ~6x faster than our CDS computers.


controls@megatron|~> speedtest
/usr/local/bin/speedtest:5: UserWarning: Module dap was already imported from None, but /usr/lib/python2.7/dist-packages is being added to sys.path
  from pkg_resources import load_entry_point
Retrieving speedtest.net configuration...
Testing from Caltech (131.215.115.189)...
Retrieving speedtest.net server list...
Selecting best server based on ping...
Hosted by Race Communications (Los Angeles, CA) [29.63 km]: 6.52 ms
Testing download speed................................................................................
Download: 6.35 Mbit/s
Testing upload speed................................................................................................
Upload: 5.10 Mbit/s
controls@megatron|~> exit
logout
Connection to megatron closed.
controls@nodus|~ > speedtest
Retrieving speedtest.net configuration...
Testing from Caltech (131.215.115.52)...
Retrieving speedtest.net server list...
Selecting best server based on ping...
Hosted by Phyber Communications (Los Angeles, CA) [29.63 km]: 2.196 ms
Testing download speed................................................................................
Download: 721.92 Mbit/s
Testing upload speed................................................................................................
Upload: 251.38 Mbit/s

Attachment 1: Screen_Shot_2017-06-04_at_1.47.47_PM.png
Screen_Shot_2017-06-04_at_1.47.47_PM.png
Attachment 2: Screen_Shot_2017-06-04_at_1.44.42_PM.png
Screen_Shot_2017-06-04_at_1.44.42_PM.png
  1067   Wed Oct 22 12:37:47 2008 josephbUpdateComputersNetwork spreadsheet
Attached in open office format as well as excel format is spreadsheet containing all the devices with IP addresses at the 40m. Please contact me with any corrections.
Attachment 1: 40m_network_10-15-08.ods
Attachment 2: 40m_network_10-15-08.xls
  7993   Mon Feb 4 15:26:10 2013 JamieUpdateComputer Scripts / ProgramsNew "getdata" program to pull NDS channel data, including test points

I've added a new program called getdata (to scripts/general/getdata) that will conveniently pull arbitrary data from an NDS server, either DQ or online (ie. testpoints).

Start times and durations may be specified.  If past data is requested, you must of course be requesting DQ channels.  If no start time is specified, data will be pulled "online", in which case you can specify testpoints.

If an output directory is specified, the retrieved data will be stored in that directory in files named after the channels.  If an output directory is not specified, no output will be

Help usage:

 

controls@pianosa:~ 0$ /opt/rtcds/caltech/c1/scripts/general/getdata --help
usage: getdata [-h] [-s START] [-d DURATION] [-o OUTDIR] channel [channel ...]

Pull online or DQ data from an NDS server. Use NDSSERVER environment variable
to specify host:port.

positional arguments:
  channel               Acquisition channel. Multiple channels may be
                        specified acquired at once.

optional arguments:
  -h, --help            show this help message and exit
  -s START, --start START
                        GPS start time. If omitted, online data will be
                        fetched. When specified must also specify duration.
  -d DURATION, --duration DURATION
                        Length of data to acquire.
  -o OUTDIR, --outdir OUTDIR
                        Output directory. Data from each channel stored as
                        '.txt'. Any existing data files will be
                        automatically overwritten.
controls@pianosa:~ 0$ 

  2892   Thu May 6 19:51:22 2010 JenneUpdatePEMNew 'Seismic Spectrum of the 40m'

For reasons unknown, the seismic spectra posted above Rosalba has been wrong since ~January when it was first posted.  The noise that we were claiming was waaaay lower than is really possible.

Rana and I checked the calibrations, and the numbers in DTT for the Ranger and the Guralp are correct (it's unknown what was being used at the time of the bad plot) - Cal for the Guralp is 3.8e-9 m/s, and for the Ranger is 1.77e-9 m/s.

Something is funny with the accelerometer calibration.  Hopefully Kevin's investigation will sort it out.  Their calibration used to be 1.2e-7 m/s^2 , but it was changed to be 7e-7 m/s^2 to match the noise level of the accelerometers with the seismometers at ~10Hz. We need to go through the calibration carefully and figure out why this is!

Posted above Rosalba for easy reference, and attached below, is the new seismic spectra.  The black trace is when the Ranger's mass is locked down, and the teal circle markers indicate the Guralp Spec-Sheet Noise Floor.

** Rana says> the y-axis in Jenne's plot is (m/s)/sqrt(Hz). The Guralp has a velocity readout bandwidth of 0.03-40 Hz, so we would have to modify the calibration to make it right in those frequencies. I believe the Ranger cal has the correct poles in it. The huge rise at low frequencies is because of the 1/f noise of the SR560.

Attachment 1: SeisRef_6May2010_AccelCalFudged.png
SeisRef_6May2010_AccelCalFudged.png
  4526   Thu Apr 14 19:05:17 2011 KojiUpdateLSCNew (temporary) LSC screen

[Jenne Koji]

The PD signals are transmitted to the suspension now.

The trigger thresholds were set to -1. This means the triggers are always on.

Attachment 1: temporary_LSC_screen.png
temporary_LSC_screen.png
  11788   Thu Nov 19 14:50:34 2015 Max IsiUpdateGeneralNew 2D histogram plot for summary pages

A new type of plot is now available for use in the summary pages, based on EricQ's 2D histogram plots (elog 11210). I have added an example of this to the SandBox tab (https://nodus.ligo.caltech.edu:30889/detcharsummary/day/20151119/sandbox/). The usage is straighforwad: the name to be used in config files is histogram2d; the first channel corresponds to the x-axis and the second one to the y-axis; the options accepted are the same as numpy.histogram2d and pyploy.pcolormesh (besides plot limits, titles, etc.). The default colormap is inferno_r and the shading is flat.

Attachment 1: C1-ALL_AB3834_HISTOGRAM2D-1131926417-86400.png
C1-ALL_AB3834_HISTOGRAM2D-1131926417-86400.png
  382   Fri Mar 14 16:56:03 2008 DmassBureaucracyComputersNew 40m control machine.
I priced out a new control machine from Dell and had Steve buy it.

GigE cards (jumbo packet capable) will be coming seperately.

Specs:
Quad core (2+GHz)
4 Gigs @ 800MHz RAM
24" LCD
low end video card (Nvidia 8300 - analog + digital output for dual head config)

No floppy drive on this one (yet?)
  1215   Sun Jan 4 13:17:23 2009 AlanOmnistructureComputer Scripts / ProgramsNew 40mWebStatus
I have set up some code in /cvs/cds/caltech/scripts/webStatus along with a cronjob on controls@nodus to generate a webStatus every half hour, at 40mWebStatus

you are welcome to add/delete lines corresponding to interesting EPICS channels, in the template /cvs/cds/caltech/scripts/webStatus/webStatus_template.html . The 2nd number is the "golden" value of the EPICS channel; it can be edited by hand, or one could copy a "golden" webStatus.html to webStatus_template.html . I think it's probably premature to automate this...

I noticed that Yoichi also has a cron job posting 40m medm screen snapshots. Very nice.

controls@nodus also runs a third cronjob, which checks if the nightly backup fails, and if so, sends an email to me.

I guess we need some kind of "official" crontab file for controls@nodus so that we know how/where to add things. So, I put one in /cvs/cds/caltech/crontab/controls@nodus.crontab
  1216   Mon Jan 5 11:21:05 2009 AlanOmnistructureComputer Scripts / ProgramsNew 40mWebStatus

Quote:

I guess we need some kind of "official" crontab file for controls@nodus so that we know how/where to add things. So, I put one in /cvs/cds/caltech/crontab/controls@nodus.crontab


Alan and I agreed that we should edit the crontab by "crontab -e" command rather than editing the "official" crontab in /cvs/cds/caltech/crontab/.
After confirming that the new crontab works as expected, you are encouraged to make a copy of the new crontab into /cvs/cds/caltech/crontab/ as a backup.
Then do "svn ci" in the directory.
  13707   Mon Mar 26 23:49:27 2018 gautamUpdateGeneralNew ADC Adaptor Board installed in C1LSC expansion chassis

Todd informed me that the ADC Timing adaptor boards we had ordered arrived today. I had to solder on the components and connectors as per the schematic, though the main labor was in soldering the high density connectors. I then proceeded to shut down all models on c1lsc (and then the FE itself). Then classic problem of all vertex machines crashing when unloading models on c1lsc happened (actually Koji noticed that this was happening even on c1ioo). Anyways this was nothing new so I decided to push ahead. 

I had to get a cable from Downs that connects the actual GS ADC card to this adaptor board. I powered off the expansion chassis, installed the adaptor board, connected it to the ADC card and restarted the expansion chassis and also the FE. I also reconnected the SCSI cable from the AA board to the adaptor card. It was a bit of a struggle to get all the models back up and running again, but everything eventually came back(after a few rounds of hard rebooting). I then edited the c1x04 and c1lsc simulink models to reflect the new path for the X arm ALS error signals. Seems to work alright.

At some point in the afternoon, I noticed a burning smell concentrated near the PSL table. Koji traced the smell down to the c1lsc expansion chassis. We immediately powered the chassis off. But Steve later informed me that he had already noticed an odd burning smell in the morning, before I had done any work at the LSC rack. Looking at the newly installed adaptor card, there wasn't any visual evidence of burning. So I decided to push ahead and try and reboot all models. Everything came back up normally eventually, see Attachment #1. Particle count in the lab seems a little higher than usual (actually, according to my midnight measurement, they are ~factor of 10 lower than Steve's 8am measurements), but Steve didn't seem to think we should read too much into this. Let's monitor the situation over the coming days, Steve should comment on the large variance seen in the particle counter output which seems to span 2 orders of magnitude depending on the time of the day the measurement is made... Also note that there is a BIO card in the C1LSC expansion chassis that is powered by a lab power supply unit. It draws 0 current, even though the label on it says otherwise. I a not sure if the observed current draw is in line with expectations.


The spare (unstuffed) adaptor cards we ordered, along with the necessary hardware to stuff them, are in the Digital FE hardware cabinet along the east arm.

Steve:  particle count in the 40m is following outside count, wind direction, weather condition .....etc. The lab particle count is NOT logged ! This is bad practice.

Attachment 1: CDS_20180326.png
CDS_20180326.png
  9197   Thu Oct 3 10:29:03 2013 masayukiUpdateGreen LockingNew ALS autolocker flowchart

 

 [Manasa, Masayuki]

We made a new flowchart of ALS autolocker. We added the additional step to find the beat note frequency. We have to find a way to read the PSL temperature. By reading the PSL temperature we can decide the sweep range for the end green laser temperature with the curve which measured in previous measurement (in this entry)

We have three thresholds of error signal. One is the threshold for checking the arms are stabilized or not. It should be some hundreds count. Another threshold is to check that the suspensions are not kicked. This should be some thousands counts (in flow chart, it is 2K counts). The other is to check the optimal servo gain. If the servo gain is too high, the UGF is also too high and we will not have enough gain margin. The error signal start to oscillate at the UGF. We will check this oscillation and find the optimal gain. In flow chart this threshold is 1K counts.

Attachment 1: flowchart.pdf
flowchart.pdf
  9865   Mon Apr 28 10:59:54 2014 KojiUpdateLSCNew ALS servo design: expected error signals

The expected error signals derived from the estimated free running error signals of the ALS.

1) Previously estimated free-running noise (blue)
2) Previous in-loop ALS error signal (red)
3) Estimated error signal with the new servo (green)
4) Out-of-loop noise of the ALS with the arm controlled with the IR PDH (black)

Now the error signal (green) is expected to be very white.
The suppressed noise between 3 to 20Hz are below the sensing noise level.
There seems a little excess at 24.5Hz and 28Hz. If it is limiting the RMS, we need to take care of them.

Attachment 1: ALSX_SPE_new.pdf
ALSX_SPE_new.pdf
Attachment 2: ALSY_SPE_new.pdf
ALSY_SPE_new.pdf
  9867   Mon Apr 28 11:08:11 2014 KojiUpdateLSCNew ALS servo design: expected error signals

Here are the MATLAB scripts and LISO codes for all of these servo analyses

Attachment 1: 140421_ALS_servo.zip
  9866   Mon Apr 28 11:03:57 2014 KojiConfigurationLSCNew ALS servo implemented for the X arm

The new ALS/LSC servo was implemented for the X arm.

I'll upload more data later but here I make quick remarks:

- We need to give the gain of 12 to have correct UGF with the ALS.

- With this servo, the Xarm PDH lock oscillates with the gain of 0.02. We need to lower the gain to 0.015.
  Also FM trigger should be changed not to trigger unused FMs (FM7/8)

  9874   Tue Apr 29 01:10:16 2014 KojiConfigurationLSCNew ALS servo implemented for the X arm

New ALS servo performance

Attachment 1:

Comparison between the old (orange) and new (red). The new error signal (red) is suppressed like a white noise as expected.

Comparison between the out-of-loop evaluation (black) and the in-loop signal (red). Below 50Hz the out-of-loop is limited by the sensor-noise like something.
This out-of-loop stability was measured with the ALS stayed at the top of the resonance and calibrated the POX11 error signal.

Attachment 2:

New ALS servo with the LSC PDH signal. When the PDH signal is used for the control, FM4 is additionally used.
In this condition, the error signal was measured and calibrated into frequncy noise (Hz/sqrtHz).

By comparing the POX (with the new servo) and POY (with the old servo) signals, one can see that the new servo has better supression below 30Hz with almost no cost at ~100Hz.

Attachment 1: ALSX_SPE.pdf
ALSX_SPE.pdf
Attachment 2: ALSX_PDH_SPE.pdf
ALSX_PDH_SPE.pdf
  10554   Tue Sep 30 17:26:18 2014 ericqUpdateLSCNew AO cable in place

I've installed a new 2pin lemo cable going from the CM servo out to in2 of the MC servo board, and removed the temporary BNC. I used some electrical tape to give the cable some thickness where the lemo head screws on to try to strain relieve the solder joints; hopefully this cable is more robust than the last. 

I put an excitation into the CM board, and saw it come out of MC_F, so I think we're set. 

  5489   Tue Sep 20 20:58:35 2011 AnamariaConfigurationLSCNew AP Table Drawing

As promised, I have made a final AP table drawing, including the MC camera relocation changes by Kiwamu. I have posted it in the wiki on the tables list, and on the AP table page I've attached the inkscape .svg I used to make it, if someone needs to do small modifications.

Attached is a pdf version of it.

Big changes:

1) REFL beam has been split into 4, to go in equal powers and equal beam size to the now 4 REFL RFPDs, 11, 33, 55 and 165. A lens had to be added for REFL165 because it's a 1mm PD instead of 2mm like the other 3.

2) MC camera has moved.

3) I've cleaned up most of the random components on the table, put them away, and tidied up the cabling.

 

Attachment 1: APtableSep20th.pdf
APtableSep20th.pdf
ELOG V3.1.3-