Month: September 2000

Intermittent, brief periods of high latency…

Sun Sep 10 15:28:33 PDT 2000 — Intermittent, brief periods of high latency and packet loss on UUNet’s San Francisco network. We are seeing fleeting amounts of high latency and packet loss on UUNet’s San Francisco network, and we are investigating the problem with UUNet. -Scott

(707) 929-9816 and 17 “not in service”.

Sun Sep 10 15:22:28 PDT 2000 — (707) 929-9816 and 17 “not in service”. Our Cobb Mountain tech support and sales lines — 929-9817 and 929-9816 — are giving a “not in service” recording. We are investigating the problem with the phone company serving that number. Note that both Cobb Mountain modem lines (929-9811 and 993-0174) are fine. -Scott, Chris, and Kevan

Excessive disk usage on /home.

Sun Sep 10 13:08:36 PDT 2000 — Excessive disk usage on /home. Between 10am this morning and 12:59 this afternoon, the “/home” filesystem was filled up — shell users saw this as a “no space left on device” error. Impact: shell users couldn’t write data to the filesystem. This includes procmail processes running on a user’s behalf that filter mail into individual files under their /home directory. (If you use procmail in this manner, your mail may have been delayed.)

After executing a “quota resize” on our Network Appliance (NetApp) to clear the condition, we found the reason for the excessive disk usage on /home: -rw——- 1 culprit user 1402852697 Sep 10 12:59 ErrorLog While 1.4 GB files are not normally allowed for members, investigation of the culprit’s shell environment reveals that he had selected options that, as a side effect, did not implement the standard file size checks for user files. This will be corrected shortly.

Additionally (and much to our chagrin), the monitoring tool that pages us when free space on NetApp filesystems shrinks too low didn’t detect the problem. So, though the tool detects physical space problems, the part that checks free space for NetApp quota trees isn’t working properly — and we didn’t know that was the case because we’ve never had a quota tree fill up while using the tool. This, too, will be corrected shortly.

Finally, we apologize to anyone who noticed the problem, most especially those who contacted tech support and got a wrong answer. We will review our processes — both computer and human — to ensure this doesn’t happen again. -Scott, Eli, and Dane

Our 1003 dial group was returning a reorder…

Sat Sep 9 18:36:31 PDT 2000 — Our 1003 dial group was returning a reorder tone for about 15 minutes; it seems that one of the T1 PRI cards lost it’s tiny mind, I’ve swapped it with the very last card in the group and a reboot seems to have fixed it. -Dane

Fiber Cut in Bay Area.

Fri Sep 8 11:34:57 PDT 2000 — Fiber Cut in Bay Area. Global Crossing (gblx.net) has experienced what has been termed a “catastrophic” fiber cut. Multiple circuits are effected, and there is the possibility that multiple ISP’s are affected. Sonic.net members may experience high latency and packet loss to gblx-connected sites while gblx’s network is repaired. According to our “Internet-weather” monitoring, neither UUNet or Cable & Wireless are affected, nor is Sonic.net. -Scott

BroadLink will be doing radio work that may…

Fri Sep 8 16:50:42 PDT 2000 — BroadLink will be doing radio work that may affect customer links this evening. Customers may experience a brief interruption or degraded service levels at 11 p.m. tonight while equipment is exchanged. Downtime is expected to be approximately fifteen minutes. If you are experiencing degraded service levels outside this time-frame please contact support@broadlink.com. -Dane and Anna

BroadLink found that a network switch had…

Thu Sep 7 14:19:32 PDT 2000 — BroadLink found that a network switch had locked up, and a reboot fixed the trouble. Total downtime for BroadLink customers was 15 minutes. They will be working with the manufacturer to see if a cause can be pinned down. There has been no previous trouble with this equipment, and it’s connected to a remote power management unit, so it can be recycled from BroadLink’s offices. -Scott, Dane, Kelsey, Eli and Shane (the doughnut guy)

BroadLink’s backhaul circuit down.

Thu Sep 7 14:09:12 PDT 2000 — BroadLink’s backhaul circuit down. BroadLink’s optical circuit to their distribution hub is currently down. BroadLink engineers are scrambling to correct the problem, and Sonic.net is assisting wherever possible. -Scott, Dane, Kelsey, Eli

A few minutes ago, scurrilous ruffians…

Mon Sep 4 17:35:07 PDT 2000 — A few minutes ago, scurrilous ruffians launched a denial-of-service attack on a Sonic.net customer’s network. The attack stopped before we could catch the perpetrators. The attack lasted about 10 minutes, and caused intermittent loss of connectivity. -Scott