Ceph Intro & Architectural Overview

Поделиться
HTML-код
  • Опубликовано: 27 июн 2015
  • Ceph is a free software storage platform designed to present object, block, and file storage from a single distributed computer cluster. Ceph's main goals are to be completely distributed without a single point of failure, scalable to the exabyte level, and freely-available. The data is replicated, making it fault tolerant. Ceph software runs on commodity hardware. The system is designed to be both self-healing and self-managing and strives to reduce both administrator and budget overhead. Ceph employs four distinct kinds of daemons:[4] Cluster monitors (ceph-mon) that keep track of active and failed cluster nodes Metadata servers (ceph-mds) that store the metadata of inodes and directories Object storage devices (ceph-osd) that actually store the content of files. Ideally, OSDs store their data on a local btrfs filesystem to leverage its built-in copy-on-write capabilities, though other local filesystems can be used instead.[5] Representational state transfer (RESTful) gateways (ceph-rgw) that expose the object storage layer as an interface compatible with Amazon S3 or OpenStack Swift APIs All of these are fully distributed, and may run on the same set of servers. Clients directly interact with all of them.[6] Ceph does striping of individual files across multiple nodes to achieve higher throughput, similarly to how RAID0 stripes partitions across multiple hard drives. Adaptive load balancing is supported whereby frequently accessed objects are replicated over more nodes.[citation needed] As of December 2014, underlying filesystems recommended for production environments are ext4 (small scale) and XFS (large scale deployments), while Btrfs and ZFS are recommended for non-production environments.[7] Object storage An architecture diagram showing the relations between components of the Ceph storage platform Ceph implements distributed object storage. Ceph’s software libraries provide client applications with direct access to the reliable autonomic distributed object store (RADOS) object-based storage system, and also provide a foundation for some of Ceph’s features, including RADOS Block Device (RBD), RADOS Gateway, and the Ceph File System. The librados software libraries provide access in C, C++, Java, Python and PHP. The RADOS Gateway also exposes the object store as a RESTful interface which can present as both native Amazon S3 and OpenStack Swift APIs. Block storage Ceph’s object storage system allows users to mount Ceph as a thinly provisioned block device. When an application writes data to Ceph using a block device, Ceph automatically stripes and replicates the data across the cluster. Ceph's RADOS Block Device (RBD) also integrates with kernel virtual machines (KVMs). Ceph RBD interfaces with the same Ceph object storage system that provides the librados interface and the CephFS file system, and it stores block device images as objects. Since RBD is built on top of librados, RBD inherits librados's capabilities, including read-only snapshots and revert to snapshot. By striping images across the cluster, Ceph improves read access performance for large block device images. The block device is supported in virtualization platforms, including Apache CloudStack, OpenStack, OpenNebula, Ganeti, and Proxmox Virtual Environment. These integrations allow administrators to use Ceph's block device as the storage for their virtual machines in these environments. File system Ceph’s file system (CephFS) runs on top of the same object storage system that provides object storage and block device interfaces. The Ceph metadata server cluster provides a service that maps the directories and file names of the file system to objects stored within RADOS clusters. The metadata server cluster can expand or contract, and it can rebalance the file system dynamically to distribute data evenly among cluster hosts. This ensures high performance and prevents heavy loads on specific hosts within the cluster.
    Slides
    www.slideshare.net/buildaclou...
  • НаукаНаука

Комментарии • 49

  • @damonmueller4709
    @damonmueller4709 5 лет назад +12

    3:51 -- I love how every time someone talks about Ceph they have a moment like this. A little Freudian slip. "This is the approach that WE favor.." --throat clear-- Sometimes instead you'll catch a little chuckle, as if the speaker is thinking to themselves how awesome Ceph is, as if they can't help themselves. (they're better than I am, they know a better way than I do and they straight up know it... and for once I actually like that. [Because I can learn from it.]) And it's almost like it's secretly a joke if you aren't already fully aware that it's the future.. I was worried about the unknown but I'm starting to gain a comfortable understanding. (I love RUclips learning.) Enough to deploy a 4 node cluster. Little clues show me that we're still only in the beginning with Ceph but I'm also pretty late to the party. Production ready but the best is still in store. I'm ready to party now! ... this video in particular helped me realize that I have 1 more node that I can add to my cluster. Time to put my resources to REAL use. Thank you.

  • @enmanuelh17
    @enmanuelh17 4 года назад +5

    Best tech presentation I seen in a long time

  • @Bigeinla
    @Bigeinla 2 года назад +6

    Great presentation. Exactly what I was looking for.

  • @suyashdongre
    @suyashdongre 8 месяцев назад

    THE best explanation of Ceph I have seen so far. Great work!

  • @BrianThomas
    @BrianThomas Год назад +7

    I've been searching all day for a decent explanation of CEPH and this is by far one of the "BEST" I have seen. Thank you very much. The light bulb just came on and it's burning bright. How can I get more of that type of training and explanation?

  • @amirmohg1319
    @amirmohg1319 Год назад +1

    determinism means that a specific key always generates a specific value no matter how many times the algorithm is rolled. This basically means there is no randomness involved in the hashing algorithm itself.

  • @harunisiaho
    @harunisiaho 6 лет назад +33

    Simple and clear explanation

    • @ewenchan1239
      @ewenchan1239 4 года назад

      ​@Peter Lustig
      Can you clarify how this isn't decentralised storage?
      For those of us that are new to this sort of stuff, what would you consider to be decentralised storage if you don't think that this is considered to be decentralised storage?

    • @ewenchan1239
      @ewenchan1239 4 года назад

      This is a nice, simple, clear, 20,000 ft overview of parallel/distributed storage.
      It would be nice to see a comparison between Ceph, Lustre, pNFS, and Gluster to see what are the differences between each of these different technologies and the relative pros and cons of each.

  • @lostsoulparty
    @lostsoulparty 4 года назад

    Some good info, looking forward to setting up my first CEPH cluster

  • @GustavoPantuza
    @GustavoPantuza 7 лет назад +5

    Great talk. Thank you.

  • @fatmaghandour9427
    @fatmaghandour9427 5 лет назад

    Thank you for posting this great talk

  • @DiegoBarrios
    @DiegoBarrios 2 года назад

    Awesome information. Thanks a lot!

  • @jarabers2320
    @jarabers2320 4 года назад

    Great explanation, easy to understand

  • @snowman4839
    @snowman4839 6 лет назад +2

    fantastic talk

  • @MistaSmith
    @MistaSmith 7 лет назад +24

    lol. how the description text below this video is better than the whole ceph website at explaining what ceph is and does.

    • @axiom1650
      @axiom1650 4 года назад +3

      Still true 3 years later

    • @SaifUlIslam-di5xv
      @SaifUlIslam-di5xv 2 года назад +2

      @@axiom1650 yeah ... the description is more detailed. I think the website has some collections of buzzwords. Good for business leaders, not really that great for engineers.

  • @RandyRandersonthefamous
    @RandyRandersonthefamous 3 года назад

    Very cool. Will be the future!

  • @user-sd2sh2wo9o
    @user-sd2sh2wo9o 8 месяцев назад

    The best intro. Thanks

  • @bunthaideng2492
    @bunthaideng2492 4 года назад

    Thank you!

  • @BBoyJamalloRock
    @BBoyJamalloRock 2 года назад

    really good stuff

  • @danielkrajnik3817
    @danielkrajnik3817 3 года назад

    yeah in overall thats a pretty comprehensive introduction to ceph, nice

  • @abhishekkr9822
    @abhishekkr9822 5 лет назад +2

    I could see some correlation between the Dynamic Subtree Partitioning which CEPH uses and the DNE (Distributed Namespace) used in Lustre. Do they branch out from the same origin?

  • @tzutalin7143
    @tzutalin7143 7 лет назад +2

    Nice talk.

  • @MohamedGamil
    @MohamedGamil 5 лет назад +1

    where can i check the slides used in this great keynote ?

  • @HenryPan
    @HenryPan 5 лет назад

    Cool

  • @MilorRamadi
    @MilorRamadi 3 месяца назад

    I think the description below the video is a part of science public paper. Have any one help me find that title or link on google scholar?

  • @afortiorama
    @afortiorama 2 года назад

    This is still a really good presentation

    • @jordanhazen7761
      @jordanhazen7761 5 месяцев назад

      The only part that seems especially dated now is mention of each OSD running on top of a normal filesystem, which is less common these days for new deployments. Be sure to read up on Bluestore vs. Filestore. Bluestore (OSD directly atop block devices, with optional separate WAL / write-ahead-log) can have much better write performance

  • @cemery50
    @cemery50 5 лет назад +1

    It remindes me of the 432 chips by Intel and distributed computing and distributed voted i/o processing.
    Does it allow for offline and non-deterministic osd in the factoring of recovery/deletion/renaming options ?
    I will look on the sites you reccommended at: ruclips.net/video/7I9uxoEhUdY/видео.html

  • @isbestlizard
    @isbestlizard 2 года назад

    This is cool. What is max performance of Ceph? Can it read/write at 50 GB/sec to a clustered fs?

  • @senthilkumarrengarajan4028
    @senthilkumarrengarajan4028 6 лет назад

    Good one. How the fragmentation is handled

    • @WileECoyotey
      @WileECoyotey 6 лет назад

      RDOS sits on top of a filesystem;

  • @richrein1
    @richrein1 4 года назад

    Are hard links supported? (not softlinks) Its required for things like Cassandra and other high bandwidth file persisted products.

    • @MagnumCarta
      @MagnumCarta 3 года назад

      Since RADOS sits on top of a pre-existing filesystem such as ext4, xfs, btrfs, etc. hard links are supported.

  • @mortenaa
    @mortenaa 5 лет назад +1

    Anyone have the slides for this talk?

    • @catborise
      @catborise 5 лет назад +2

      www.slideshare.net/buildacloud/ceph-intro-and-architectural-overview-by-ross-turk

  • @horizonbrave1533
    @horizonbrave1533 5 лет назад

    Great talk...but Now I'm baffled as to how CEPH works with the other Openstack storage solutions like Swift and Cinder...?

    • @MagnumCarta
      @MagnumCarta 3 года назад

      From what I gathered on the Openstack documentation, Cinder allows a driver to be provided for how to write to the disks. Cinder will manage control of the disks while the Ceph RBD driver will manage how to communicate with the drives by managing a storage pool. With Cinder you can provide a batch of drives to be the target for the storage pool that Ceph will manage. In the following hyperlink there is an example given for a scenario where you might want to create three tiers of SSDs: "Gold", "Silver", and "Bronze" respectively that would correspond to the speed of the SSDs and number of replications made for backups.
      superuser.openstack.org/articles/ceph-as-storage-for-openstack/#:~:text=Cinder%20is%20the%20block%20storage,as%20gold%2C%20silver%20or%20bronze.

  • @isbestlizard
    @isbestlizard 2 года назад

    Huh. File systems should be fully deterministic. The same data in the same file should put it in the same blocks. That would allow reduction again.

  • @vcool
    @vcool 7 лет назад

    This is the same old 2012 talk.

  • @balla2172
    @balla2172 5 лет назад

    only issue i see here is as you scale out you scale failure points instead of 1 huge storage appliance with redundancy you have tons of small appliances with no redundancy

  • @raidenwondershareworldwind6447
    @raidenwondershareworldwind6447 5 лет назад

    vox news opt/ceph intro&archtectural overview

  • @bunthaideng2492
    @bunthaideng2492 4 года назад

    I cannot find ceph.com/qsg

  • @smorfnimda
    @smorfnimda 11 месяцев назад

    Why can’t anyone explain CEPH without the BLOATED explanation??? WTF? Can’t you just go step by step on how to install it and configure it and that’s it??? I’m a web Developer and when I explain PHP or JAVA I don’t explain HOW they work, I explain how to get it going!

  • @parhamzardoshti9675
    @parhamzardoshti9675 3 года назад

    Great talk. Thank you.

  • @danielkogans
    @danielkogans 3 года назад

    Thank you!