arls1481's picture

QLogic SanBox 9100 question


I've got a Sanbox 9100 with (5) SB9004 - 4/2/1G blades in it and have run out of ports. I'm having a hard time sourcing out 4/2/1G blades but can come up with 8/4/2G SB9008 blades pretty easily. Anyone know if its kosher to pop those newer blades in to the 9100? I'm assuming it will work fine and that any ports I connect that have 8G at both ends will utilize it and it downscales otherwise???

bartbroek's picture

Looking for second hand Active Raid storage


I'm looking for a ActiveRaid main chassis (with or without drives) or expansions chassis to replace an Expansion chassis that is giving me trouble. I'm also looking for some spare Active raid controllers.

We are a post-production company based in Amsterdam, The Netherlands.

Thanks in advance for any suggestions!


messenger82's picture

Question about unsupported volume resizing

This something wholly unsupported but we're playing around with it anyway for experimental reason. If anyone has a thoughts I would love to hear it.

We have a lab / sand box system using commodity storage. Specifically it's a FreeNAS server. We are using it as an iSCSI target and have setup a series of zvols as device extents. On the client side we have Mac pro's, we use one physical port for the iSCSI network, the other port is virutalized with a connection to our public network and metadata network. We are using globalsan client side to connect to the iSCSI storage. We labeled the LUNs, XSAN let us use them to create a SAN.

So my question, one of the things were playing with is resizing volumes and adjusting storage on the fly (or at least with a restart). So lets say I re-size my zvols on my iSCSI side (which I did) which should create a larger volume on the XSAN side, but it doesn't. The volume stays the same size. So how would I tell XSAN that my LUNs (and subsequently my volume) should be larger?

Yes, this is like crossing the streams with our proton packs. It's just a fun thought experiment about using low cost hardware to take advantage of FCPX's SAN Location benefits in low cost labs.

MattZ's picture

Qlogic 6 switch Stack


I am wondering how many setups out there have 6 switches in the fibre stack?

I am again seeing a build out with 6 switches in the stack, and in the past i have seen huge issues with latency
so wanted feedback on users that have had success with that many switches?
I know a 9000 series is the smart way to go.

MDCs both in standby, volume won't mount

Had some serious funkiness with our SAN volume. Lots of inode issues. I ran cvfsck -j, then -nv, then -wv, which seems to have cleaned up all the errors. After trying to start the volume again, it immediately went into a RPL_upgrade process. A couple times, that failed near the end (fsm process crashed), then the volume would fail over to the other MDC and start the who RPL process again. I saw a "completed" message at least three times now, but the volumes will still not mount correctly. Both MDCs have their fsm processes running in standby mode now, though the volume appears to be started and hosted by the master in Xsan Admin. Mounting the volume anywhere just causes another crash.

After running cvadmin activate on the master MDC, it eventually tells me:

Admin Tap Connection to FSM failed

This also kicks off another round of RPL upgrades.

I'm hoping I don't have to completely restore our backup, because that'll take days… Any suggestions great appreciated. Thank you!

hernand's picture

Xsan display incorrect Lun size

I have these problem. The Xsan display incorrect lun size. As you can see the Disk Utility app shows the correct configuration, but the Xsan admin displays one LUN of 20TB (I dont where it came from) and another 42TB (XsanLUN1).
I think it has to show 2 LUNS of 20TB each acording to Disk Utility app shows.
The drive began experience read write problems and I think it would be parte of these problem.
Can you help me?




thomasb's picture

Qlogic getting out of the Fibre Channel switch business


Did anybody else notice these news?

QLogic's save-the-biz pitch: Ditch the switch glitch - get rich
Server adapter firm to focus on, er, server adapters...

We've had three Qlogic 5800-switches "bricked" in under a year at work, running firmware Two of them because of a power failure, and one of them simply didn't want to turn on again after turning it off. But isn't a power failure simmilar to turning it off? I know there is a shutdown command in the command line interface, but this is very well hidden in some old manuals for the SANbox switches. We have always simply cut the power to turn off Qlogic 5600/5800 switches, without issues, and suddenly this year three 5800 switches die when cutting the power. Not good.

Anybody else had issues with 5800-switches going into "brick mode" with the heartbeat LED constantly lit? Pushing the maintenance mode button doesn't help. It just tries to reload, and then it's stuck again.

The 5600-swtiches we have, have been running solid since 2006 though, without any serious issues.

Time to move on I guess.

ygini's picture

AFP and Barracuda LB 340, not stable?



I've a Xsan setup with two 10.8 server used for AFP (and for few systems, SMB) and I try to use a Barracuda LB 340 to handle HA constraints.

We don't have a need of big bandwidth, but we need easy to use HA system for end user.

The LAN is a flat network, so I've configure the LB 340 with the TCP proxy mode.

Most of the time, it work. But users connected to the AFP share through the LB experience some problems. File in read only, impossible to copy file from a share to an other, etc.

Same users, but connected directly to one of the OS X Server don't have the problem.

So, I'm wondering. How do you handle your HA setup, and if you use Barracuda products, how do you configure it?

matx's picture

Promise SANLink2

Just noticed that Promise has released information about their Thunderbolt 2 based products, the SANLink2 fibre channel adapter and their Pegasus2 RAIDs.

Not sure when are released to market officially. But we finally have a dual 8GB thunderbolt to fibre channel adapter from Promise.

Xsan AFP reshare problems

Hi all,

Firstly, sorry for cross-posting, but I did already post this over on [url=]Creative Cow/url.

I've been hitting a huge problem with our Xsan reshare this week. I'm getting to the end of my rope pretty fast, so any insights are greatly appreciated.

Our setup:

Xsan 2.3
2x MDCs @ 10.7.5
4x (ActiveStorage) head raids + 1x head chassis for metadata and other data
Qlogic 5600 switches
1x beefy Mac Pro AFP reshare @ 10.7.5
- 12-core @ 2.4GHz
- 64GB RAM
- 6x 4Gb FC link
- 2x 10Gb Ethernet bonded link via SmallTree card (lab1)
- 6x 1Gb Ethernet bonded link via SmallTree card (lab2)
Lab 1 = 40x 2013 iMacs @ 10.8.4
Lab 2 = 17x 2012 iMacs @ 10.8.4

The media that the students are using is primarily old DV-NTSC footage or AVHCD right off our new cameras. We also have several students building simple websites, and they store all their source files and images on this volume. The volume is definitely able to handle these data rates, so the problem must be with my reshare configuration. After a server reboot, the labs all perform fine, but as the day wears on, we eventually reach a point where new users can no longer connect to the AFP reshare. One class arrives, logs in, edits, logs out, (repeat), but by the fourth class, the students get a nasty prompt that the sever is unavailable. Users who are already connected can continue working just fine.

In additional to that, we occasionally have users experience choppy connections to the server. Their footage will suddenly just pause for about 30 seconds. A reboot of the client workstation usually clears this up, but not always (it'll crop up again). When I check the AFP connection list on the server, I usually see two or more duplicate connections from the workstation in question, the older "stale" ones all with increasing idle times while the active connection shows 0 idle time.

When new connections can no longer be made to the server, that's usually the point where I need to run and forcibly shut it off. Even after a fresh reboot of the server and the clients, things aren't perfect. I'm now seeing constant metadata traffic between my reshare server and the acting MDC. It's all very small data (<1MB/s), but it's constant, even when no users are connected via AFP. My acting MDC also has >100% CPU load for the fsm process, which currently boasts 299 threads (is this normal?). The kernel process on the reshare server is constantly at about 20% CPU, and the AppleFileServer process always has about 204 threads (seems high, right?).

Some frequent log messages I'm seeing on the reshare server are:

Oct 24 18:10:39 reshare-server kernel[0]: add_fsevent: unable to get path for vp 0xffffff80881e67c0 (-UNKNOWN-FILE; ret 22; type 2)
Oct 24 18:10:39 reshare-server kernel[0]: add_fsevent: unable to get path for vp 0xffffff80881e67c0 (-UNKNOWN-FILE; ret 22; type 4)
Oct 24 19:03:51: --- last message repeated 1 time ---

Oct 24 19:55:23 reshare-server AppleFileServer[958]: received message with invalid client_id 1583
Oct 24 19:55:25 reshare-server AppleFileServer[958]: _Assert: /SourceCache/afpserver/afpserver-585.7/afpserver/AFPRequest.cpp, 2006
Oct 24 19:55:25: --- last message repeated 2 times ---

Oct 24 20:33:01 reshare-server AppleFileServer[958]: MDSChannelPeerCreate: (os/kern) invalid argument
Oct 24 20:33:51: --- last message repeated 1 time ---

The first group of log messages above seems to relate to when someone has experienced a bad (stale) connection, though I can't confirm this. After said client reboots, these message cease. I was really happy when I made it the whole morning without seeing any of them, but then they started cropping up again. All these messages tend to increase with more use of the client systems, but I can't pinpoint the cause.

I've been debating converting this setup to NFS instead of AFP, but several of our workflows rely pretty heavily on the ACLs that are in place on this volume. My understanding is that Apple's nfsd doesn't pass the ACLs along to the clients. We tried NFS for a small subset of machines for a semester, but AFP was much easier and gave us our ACLs. I also have another almost identical reshare server ready and waiting to go in the rack, but I'd like to make sure I'm heading in the right direction before I throw more hardware at the situation.

Thank you!


Subscribe to Xsanity RSS