tonyswu's picture

fsmpm Not Running on Newly Added Client

Hi,

We have a small Xsan with 2 meta data controllers and 3 clients. All machines are running Lion, and using the built-in Xsan that comes with Lion. We are trying to add a new client, but after adding the client Xsan Admin simply says fsmpm not running, and on the client when trying to enable Xsan it displays an error message saying unable to read volume list. We have wiped the client machine and re-tried multiple times with no positive result.

I took a look at the log, here are the logs from client:

8/6/12 6:43:05.644 PM fsmpm: GetCctlsFromNss failed: [errno 54]: Connection reset by peer
8/6/12 6:43:05.644 PM fsmpm: PortMapper: central control init failed!
8/6/12 6:43:05.000 PM kernel: Xsan Client Revision 3.5.0 Build 7443 Branch (457) Built for Darwin 11.0 i386 Created on Sat Jun 18 13:12:27 PDT 2011
8/6/12 6:43:06.577 PM xsand: fsmpm exited unexpectedly (exit code = 1)
8/6/12 6:43:06.592 PM xsand: fsm shutdown after 0 seconds

And here are the logs from the primary controller:

8/6/12 6:39:53.138 PM fsmpm: PortMapper: new_input authentication protocol error(39)...
8/6/12 6:41:16.820 PM xsand: unable to open '/Library/Preferences/com.apple.xsan.controller.plist' for reading: No such file or directory
8/6/12 6:41:16.944 PM Xsan Admin: ERROR: Error mounting volume…: The operation couldn’t be completed. (SANTransactionErrorDomain error 100007.) (100007)
8/6/12 6:41:17.020 PM fsmpm: PortMapper: launching configuration reload thread
8/6/12 6:41:17.020 PM fsmpm: PortMapper: fsmpm configuration reload initiated (flags FFFFFFFF)
8/6/12 6:41:17.020 PM fsmpm: PortMapper: fsmpm configuration reload complete
8/6/12 6:41:17.020 PM fsmpm: Portmapper: File System RAS events undeliverable to Coordinator '192.168.1.250'. Please upgrade Xsan on this host.
8/6/12 6:43:05.649 PM fsmpm: PortMapper: new_input authentication protocol error(39)...

Looks like some sort of authentication error? I am stumped here, any help would be greatly appreciated.

Thanks.

seb's picture

x86 benchmarking tools

Hello,

Any suggestion for a benchmarking tool that will work with Lion ? All the tools I have used it the past (frametest, diskwriggler, xsan tuner) were power pc builds.

AJA and Black Magic utility don't support multiple streams, which is what I need...

Thanks !

JSeliski's picture

Xraid help anywhere?

I need to squeak another year out of my Xraid - I have a bunch of spare parts and drives - so here I go. However I had one half of my Xraid (Raid 50 - 14 400mb drives) all of a sudden lose the green lights on one side. Swapping out controllers isolated it as a controller problem. Replaced controller - then lost one drive at the same time. Popped a new one in and rebuilt over the next day or two. NOw it says all is fine - In Raid Admin as well - all green. But the Xraid keeps cycling through the drives with the activity indicator spiking (and drives cycling) every few seconds e. No problem in accessing drives or data Here is what it is doing
http://www.youtube.com/watch?v=ONl03MGY_cE

help ? ? thanks!

Xsan: Volumes may not mount at startup if no user logs in (Apple KB)

Under some circumstances, Xsan volumes configured to mount automatically at startup may not mount until a user has logged in.  Volumes mount correctly once a user logs in to the affected client.

Read more: http://support.apple.com/kb/TS4292

wrstuden's picture

Xsan makes the pages of the New York Times

It has taken seven years, but Xsan has made the pages of the New York Times.

In tech columnist David Pogue's post on Mountain Lion, Xsan got mentioned!

jw's picture

XRaid - Failed 1+0 - How to rebuild?

Forums: 

I have an older Apple XRaid with 14x 750GB EIDE drives. The RAID is set up with drives 1-7 in a RAID 0 stripe and drives 8-14 in a second RAID 0 stripe, then the two stripes are mirrored.

Drive 9 has failed and has been replaced. During this time the data has remained accessible, exactly as planned. Replacing the drive physically did not cause this ideal RAID configuration to be reestablished, however.

I could use a hand/verification in getting this back properly configured with mirrored stripes.

Currently, Disk Utility shows the following (all colors accurate as shown)
5.25 TB APPLE Xserve RAID Media
__[=gray]RAID Slice for "Storage"/color
5.25 TB APPLE Xserver RAID Media
__[=red]RAID Slice for "Storage"/color
[=orange]5.25 TB Storage/color
__Storage

Further, it indicates that the Mirrored RAID Set is "[=orange]Degraded/color" and the second 5.25 TB RAID Slice (disk1s2) is "[=red]Failed/color"

So far, so good - it looks like it simply did not rebuild the setup and it needs to be recreated manually - not unexpected from a stripe failure.

RAID Admin shows the following:

  • Array 1 is Online, RAID Level: RAID 0

__- All drives are Status OK

  • Array 2 is Offline, RAID Level: RAID 0

__- Drive 9 (the original failed drive) is Status: OK, Type:Spare, but does not have a [2] on it, indicating that it is not part of the array

At this pint I have poked around in RAID Admin's Utilities and Advanced sections, but I am not familiar enough with RAID Admin specifically to feel comfortable enough to actually try functions and overcome some of the scary-sounding warnings :wink:

A couple more symptoms:

  • RAID Admin, Advanced, Expand Array allows me to go through the motions of adding Drive 9 back into Array 2, but it says that it cannot expand array because it is offline or degraded.
  • RAID Admin, Utilities, Make Drive Available For Use shows that drive 9 is indeed already available for use.
  • Nowhere that I could find in RAID Admin does it appear to make any reference to the mirroring. I assume this is happening in Disk Utility?

It [u]looks/u like what I need to do is to delete Array 2 and recreate it, then recreate the mirror.

So, the question is, what do I need to do in order to recreate this RAID setup - both the stripe array and then the mirror?

Also, will this require downtime or merely cause poor performance while it rebuilds the mirror?/u

Mandakini's picture

Xsan 2.2.1- Volumes do not get mount

Hi All,

Need a urgent and quick help for the following.

I am trying to mount volumes on apple client using Xsan 2.2.1

When I hit cvadmin command I get the following error.

'The Xsan file system services on 127.0.0.1 may be stopped
Xsan administrator
Error in getting central control info'.

I have already tried the below steps many times.
1) Unistalled and installed Xsan admin 2.2.1
2) Recreated the xsan file,the config.plist file and the automount.plist file
3) Tried multiple reboots.
4) Removed the uuid file.
5) There is no .auth_secret file that needs to be removed.

Any help will be greatly appreciated.

huntson's picture

Interesting failover situation

Any clue why my volumes are failing over when there appears to be no issue? I have my switch configured correctly with fresh installs of Lion and the latest updates. Here is my log. I am seeing a few errors which don't necessarily make sense as everything is properly communicating and all.

] 0x103581000 INFO Disk rescan found 9 disks
[0721 23:35:28] 0x103581000 NOTICE compare_disks: label transition for disk /dev/rdisk12
[0721 23:35:41] 0x102e81000 NOTICE PortMapper: waking up diskscan thread.
[0721 23:35:41] 0x103581000 INFO Starting Disk rescan
[0721 23:35:41] 0x103581000 INFO Disk rescan delay completed
[0721 23:35:41] 0x103581000 INFO Disk rescan found 10 disks
[0721 23:35:41] 0x103581000 NOTICE compare_disks: label transition for disk /dev/rdisk6
[0721 23:36:49] 0x7fff738b8960 NOTICE PortMapper: launching configuration reload thread
[0721 23:36:49] 0x1003af000 NOTICE PortMapper: fsmpm configuration reload initiated (flags FFFFFFFF)
[0721 23:36:49] 0x1003af000 INFO NSS: Primary Name Server is '192.168.0.5' (192.168.0.5)
[0721 23:36:49] 0x10370a000 INFO NSS: Name Server '192.168.0.5' (192.168.0.5) port is 49858, revision is 0x0102.
[0721 23:36:49] 0x1003af000 INFO NSS: Secondary #1 Name Server is '192.168.0.6' (192.168.0.6)
[0721 23:36:49] 0x102f87000 (debug) nss_port_acquire_thread [4344803328] exit
[0721 23:36:49] 0x1003af000 (debug) FSS 'San' STOPPED -> STOPPED (idle)
[0721 23:36:49] 0x1003af000 NOTICE PortMapper: fsmpm configuration reload complete
[0721 23:36:49] 0x10378d000 INFO NSS: Name Server '192.168.0.6' (192.168.0.6) port is 51531, revision is 0x0102.
[0721 23:36:49] 0x103281000 (debug) nss_port_acquire_thread [4347924480] exit
[0721 23:36:49] 0x7fff738b8960 (debug) Start: Setting AUTOSTART for FSS 'San'
[0721 23:36:49] 0x1003af000 NOTICE PortMapper: Starting FSS service 'San[1]' on mdc.xsan.rvchost.int.
[0721 23:36:49] 0x7fff738b8960 (debug) FSS 'San' STOPPED (idle) -> LAUNCHED, next event in 60s
[0721 23:36:49] 0x7fff738b8960 (debug) FSS 'San' LAUNCHED -> REGISTERED
[0721 23:36:49] 0x7fff738b8960 NOTICE PortMapper: FSS 'San'[1] (pid 436) at port 49230 is registered.
[0721 23:36:49] 0x7fff738b8960 (debug) Dropping 192.168.0.5 coordinator 0 for new 49858
[0721 23:36:49] 0x7fff738b8960 INFO NSS: Standby FSS 'San[1]' at id 192.168.0.5 port 49230 (pid 436) - registered.
[0721 23:36:49] 0x103687000 NOTICE Portmapper: File System RAS events undeliverable to Coordinator '192.168.0.5'. Please upgrade Xsan on this host.
[0721 23:36:49] 0x7fff738b8960 (debug) Dropping 192.168.0.6 coordinator 0 for new 51531
[0721 23:36:49] 0x7fff738b8960 INFO NSS: Standby FSS 'San[0]' at id 192.168.0.6 port 49198 (pid 426) - registered.
[0721 23:36:49] 0x103687000 NOTICE Portmapper: File System RAS events undeliverable to Coordinator '192.168.0.6'. Please upgrade Xsan on this host.
[0721 23:36:49] 0x7fff738b8960 (debug) NSS: Coordinator 192.168.0.5 flags changed from 0x2 to 0x7
[0721 23:36:49] 0x7fff738b8960 (debug) NSS: Coordinator 192.168.0.5 id is 192.168.0.5
[0721 23:36:49] 0x7fff738b8960 (debug) Heartbeat from ID 192.168.0.5 updating LOCAL San to 192.168.0.5:49230
[0721 23:36:49] 0x7fff738b8960 (debug) NSS: Coordinator 192.168.0.6 flags changed from 0x2 to 0x7
[0721 23:36:49] 0x7fff738b8960 (debug) NSS: Coordinator 192.168.0.6 id is 192.168.0.6
[0721 23:36:49] 0x7fff738b8960 (debug) NSS: Computing nss_coord_sum
[0721 23:36:50] 0x7fff738b8960 INFO NSS: election initiated by 192.168.0.6:51531 (id 192.168.0.6) - admin request.
[0721 23:36:50] 0x7fff738b8960 INFO NSS: Vote call for FSS San is inhibited - vote dis-allowed.
[0721 23:36:56] 0x7fff738b8960 (debug) NSS: FSS mount list for client 192.168.0.6 (id 192.168.0.6) - San
[0721 23:36:56] 0x7fff738b8960 (debug) NSS: New mount registered for 'San'.
[0721 23:36:56] 0x7fff738b8960 (debug) NSS: FSS mount list for client 192.168.0.5 (id 192.168.0.5) - San
[0721 23:36:57] 0x102f87000 NOTICE PortMapper: Mount Event for /Volumes/San on /dev/disk13
[0721 23:37:02] 0x103581000 INFO Starting Disk rescan
[0721 23:37:41] 0x103581000 INFO Disk rescan delay completed
[0721 23:37:41] 0x103581000 INFO Disk rescan found 10 disks
[0721 23:40:14] 0x7fff738b8960 (debug) find_fsm fsm San ipaddr 192.168.0.6 port 49198 TestLink failed: getsockopt(SO_ERROR) returned error 61 [errno 61]: Connection refused
[0721 23:40:14] 0x7fff738b8960 INFO NSS: Active FSS 'San[0]' at 192.168.0.6:49198 (pid 426) - dropped.
[0721 23:40:15] 0x7fff738b8960 INFO NSS: election initiated by 192.168.0.6:51531 (id 192.168.0.6) - client request.
[0721 23:40:15] 0x7fff738b8960 (debug) find_fsm fsm San ipaddr 192.168.0.6 port 49198 TestLink failed: getsockopt(SO_ERROR) returned error 61 [errno 61]: Connection refused
[0721 23:40:15] 0x7fff738b8960 NOTICE PortMapper: Initiating activation vote for FSS 'San'.
[0721 23:40:15] 0x7fff738b8960 (debug) Initiate_nss_vote for FSS San
[0721 23:40:15] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.5' (192.168.0.5:49858).
[0721 23:40:15] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.6' (192.168.0.6:51531).
[0721 23:40:15] 0x7fff738b8960 (debug) NSS: FSS activation initiated by coordinator 192.168.0.6:51531 (id 192.168.0.6) votes 1
[0721 23:40:15] 0x7fff738b8960 INFO NSS: Vote call for FSS San is inhibited - vote dis-allowed.
[0721 23:40:20] 0x103281000 NOTICE PortMapper: Reconnect Event for /Volumes/San
[0721 23:40:20] 0x103281000 NOTICE PortMapper: Requesting MDS recycle of /Volumes/San
[0721 23:40:26] 0x7fff738b8960 INFO NSS: Standby FSS 'San[0]' at id 192.168.0.6 port 49365 (pid 510) - registered.
[0721 23:40:50] 0x7fff738b8960 NOTICE PortMapper: Stopping FSS 'San'
[0721 23:40:50] 0x7fff738b8960 NOTICE PortMapper: FSS 'San' has been stopped.
[0721 23:40:50] 0x7fff738b8960 (debug) FSS 'San' REGISTERED -> DYING, next event in 60s
[0721 23:40:50] 0x7fff738b8960 INFO NSS: Standby FSS 'San[0]' at 192.168.0.6:49365 (pid 510) - dropped.
[0721 23:40:51] 0x7fff738b8960 NOTICE PortMapper: Initiating activation vote for FSS 'San'.
[0721 23:40:51] 0x7fff738b8960 (debug) Initiate_nss_vote for FSS San
[0721 23:40:51] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.5' (192.168.0.5:49858).
[0721 23:40:51] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.6' (192.168.0.6:51531).
[0721 23:40:51] 0x7fff738b8960 INFO NSS: election initiated by 192.168.0.5:49858 (id 192.168.0.5) - client request.
[0721 23:40:51] 0x7fff738b8960 INFO NSS: Active FSS 'San[1]' at 192.168.0.5:49230 (pid 436) - dropped.
[0721 23:40:51] 0x7fff738b8960 (debug) NSS_VOTE2 to 192.168.0.5:49858
[0721 23:40:51] 0x7fff738b8960 (debug) NSS: removing vote inhibitor for FSS 'San'.
[0721 23:40:51] 0x7fff738b8960 (debug) start_fss_vote could not find FSS San in master - vote aborted.
[0721 23:40:51] 0x1003af000 (debug) Portmapper: FSS 'San' (pid 436) exited with status 0 (normal)
[0721 23:40:51] 0x1003af000 (debug) FSS 'San' DYING -> STOPPED (explicit request)
[0721 23:40:52] 0x7fff738b8960 (debug) Start: Setting AUTOSTART for FSS 'San'
[0721 23:40:52] 0x1003af000 NOTICE PortMapper: Starting FSS service 'San[1]' on mdc.xsan.rvchost.int.
[0721 23:40:52] 0x7fff738b8960 (debug) FSS 'San' STOPPED (explicit request) -> LAUNCHED, next event in 60s
[0721 23:40:52] 0x7fff738b8960 (debug) FSS 'San' LAUNCHED -> REGISTERED
[0721 23:40:52] 0x7fff738b8960 NOTICE PortMapper: FSS 'San'[1] (pid 520) at port 49538 is registered.
[0721 23:40:52] 0x7fff738b8960 INFO NSS: Standby FSS 'San[1]' at id 192.168.0.5 port 49538 (pid 520) - registered.
[0721 23:40:52] 0x7fff738b8960 (debug) Heartbeat from ID 192.168.0.5 updating LOCAL San to 192.168.0.5:49538
[0721 23:40:52] 0x7fff738b8960 INFO NSS: Standby FSS 'San[0]' at id 192.168.0.6 port 49377 (pid 512) - registered.
[0721 23:40:53] 0x7fff738b8960 NOTICE PortMapper: Initiating activation vote for FSS 'San'.
[0721 23:40:53] 0x7fff738b8960 (debug) Initiate_nss_vote for FSS San
[0721 23:40:53] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.5' (192.168.0.5:49858).
[0721 23:40:53] 0x7fff738b8960 (debug) NSS: sending message (type 2) to Name Server '192.168.0.6' (192.168.0.6:51531).

cmonico's picture

Xsan not starting after adding storage

I added storage to my Xsan this am and now I cant get it started...I get this error in terminal:

State: BLOCKED (configuration file problem) 2012-07-21 09:44:49
Last Admin: START 2012-07-21 09:44:48
Last Termination: exit(19) 2012-07-21 09:44:49
Launches 5, core dumps 0, flags

FSS 'XSAN01' start unsuccessful. Check event log for more details.

Any advice would be greatly appreciated.

CitraMalus's picture

Xsan 1.4 LUN got unlabeled-Volume does not mount -SOLVED

Hey Guys..

One of our client tried something terrible on a live Xsan Volume. They connected one Windows 2000 server to FC switch and tried to merge three LUNS of the XSAN Volume from Windows OS. After failing to do so, the three XSAN LUNS got unlabeled and now the Volume does not mount. So my doubt is,

1. Can we just relabel the LUNS to mount the Volume and preserve the Data?
2. If the answer for the above is NO, then will they lose data from the entire Volume, if we try to add (Just like expanding) the LUNs after relabeling it?

Please Help. the client is a TV Broadcaster and want their precious data.
Forgot to mention, the Xsan is 1.4 and the Mac OS X 10.5.

THANKS.

Pages

Subscribe to Xsanity RSS