[prev in list] [next in list] [prev in thread] [next in thread] 

List:       ceph-devel
Subject:    Re: Able to mount some but not all nodes
From:       Gregory Farnum <gregf () hq ! newdream ! net>
Date:       2011-05-28 21:12:50
Message-ID: BANLkTims=Q=wC-2wO-UcnpsaJT-Dej-ePQ () mail ! gmail ! com
[Download RAW message or body]

Mounting a Ceph cluster isn't quite like mounting an NFS machine. You
mount by connecting to one of the monitor servers, and then you
dynamically connect and disconnect to the other machines as needed. :)
So if you can only connect to your first server, using the monitor
address, then that's exactly how it should be!
-Greg

On Sat, May 28, 2011 at 2:02 PM, Adam Yee <adamjyee@gmail.com> wrote:
> Hi,
>
> I’m new to Ceph.  I’ve installed my test cluster on Ubuntu 10.10 servers.  After
> following the instructions found on the wiki site and I’ve been able to mount
> server1 (mds0 and mon0) and server2 (osd0) like so
>
> $ sudo mount -t ceph server1:/ /mnt/ceph
>
> However, server3 (osd1) and server4 (osd2) give me the following
>
> $ sudo mount -t ceph server1:/ /mnt/ceph
> mount error 5 = Input/output error
>
> Where do I look inside the log files for mounting info?  Am I missing
> configurations that activate important debug messages in the log files?  Below
> are my configurations and output of dmesg after attempting to mount at server3.
>
> Thank you very much!
> Adam
>
> $ sudo dmesg
> ...
> [  402.180998] ceph: loaded (mon/mds/osd proto 15/32/24, osdmap 5/5 5/5)
> [  402.183563] ceph: client4101 fsid 6beefbba-0776-d6fd-003d-af78ac07b9da
> [  402.183662] ceph: mon0 10.10.207.156:6789 session established
> [138449.435257] ceph: client4105 fsid 6beefbba-0776-d6fd-003d-af78ac07b9da
> [138449.435352] ceph: mon0 10.10.207.156:6789 session established
> [138618.776870] INFO: task cosd:1948 blocked for more than 120 seconds.
> [138618.776890] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this
> message.
> [138618.776911] cosd          D 0000000100d3237c     0  1948      1 0x00000000
> [138618.776918]  ffff88013ab2be00 0000000000000086 ffffffff00000000
> 0000000000015980
> [138618.776925]  ffff88013ab2bfd8 0000000000015980 ffff88013ab2bfd8
> ffff88013a525b80
> [138618.776931]  0000000000015980 0000000000015980 ffff88013ab2bfd8
> 0000000000015980
> [138618.776937] Call Trace:
> [138618.776947]  [<ffffffff81048e89>] ? __wake_up_common+0x59/0x90
> [138618.776955]  [<ffffffff8159e5ed>] rwsem_down_failed_common+0x9d/0x210
> [138618.776962]  [<ffffffff8117a810>] ? sync_one_sb+0x0/0x30
> [138618.776966]  [<ffffffff8117a810>] ? sync_one_sb+0x0/0x30
> [138618.776972]  [<ffffffff8159e7b6>] rwsem_down_read_failed+0x26/0x30
> [138618.776979]  [<ffffffff812c1224>] call_rwsem_down_read_failed+0x14/0x30
> [138618.776985]  [<ffffffff8159da27>] ? down_read+0x17/0x20
> [138618.776991]  [<ffffffff8115545b>] iterate_supers+0x7b/0xd0
> [138618.776996]  [<ffffffff8117a86f>] sys_sync+0x2f/0x70
> [138618.777004]  [<ffffffff8100a0f2>] system_call_fastpath+0x16/0x1b
> [140246.990289] ceph: client4301 fsid 6beefbba-0776-d6fd-003d-af78ac07b9da
> [140246.990390] ceph: mon0 10.10.207.156:6789 session established
> $
>
>
> /etc/hosts
> 127.0.0.1       localhost
>
> 10.10.207.156   server1
> 10.10.207.150   server2
> 10.10.206.102   server3
> 10.10.207.44    server4
>
> # server1 - MDS and Monitor
> # server2, server3, server4 - OSD
>
> # The following lines are desirable for IPv6 capable hosts
> ::1     localhost ip6-localhost ip6-loopback
> fe00::0 ip6-localnet
> ff00::0 ip6-mcastprefix
> ff02::1 ip6-allnodes
> ff02::2 ip6-allrouters
>
>
> /etc/ceph/ceph.conf
> [global]
>       pid file = /var/run/ceph/$name.pid
>       debug ms = 1
>
> ; Monitors
> [mon]
>       mon data = /data-storage/ceph/mon$id
>       debug mon = 20
> [mon.0]
>       host = server1
>       mon addr = 10.10.207.156:6789
>
> ; Metadata servers
> [mds]
>       debug mds = 20
> [mds.0]
>       host = server1
>
> ; OSDs
> [osd]
>       debug osd = 20
> ;       sudo = true
>       osd data = /data-storage/ceph/osd$id
> ;       osd journal = /journal
>       osd journal size = 512
>       filestore journal writeahead = true
> [osd.0]
>       host = server2
>       osd addr = 10.10.207.150
> [osd.1]
>       host = server3
>       osd addr = 10.10.206.102
> [osd.2]
>       host = server4
>       osd addr = 10.10.207.204
> [mount]
>       allow = %everyone
>
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
[prev in list] [next in list] [prev in thread] [next in thread] 

Configure | About | News | Add a list | Sponsored by KoreLogic