[prev in list] [next in list] [prev in thread] [next in thread] 

List:       cassandra-user
Subject:    AWS ephemeral instances + backup
From:       Carl Mueller <carl.mueller () smartthings ! com ! INVALID>
Date:       2019-12-05 20:21:22
Message-ID: CAFjxgDy6H0wc9VkyxnbyUfq1dvXHpA55jxGmrg5LCdeXn8GZAw () mail ! gmail ! com
[Download RAW message or body]

Does anyone have experience tooling written to support this strategy:

Use case: run cassandra on i3 instances on ephemerals but synchronize the
sstables and commitlog files to the cheapest EBS volume type (those have
bad IOPS but decent enough throughput)

On node replace, the startup script for the node, back-copies the sstables
and commitlog state from the EBS to the ephemeral.

As can be seen:
https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/EBSVolumeTypes.html

the (presumably) spinning rust tops out at 2375 MB/sec (using multiple EBS
volumes presumably) that would incur about a ten minute delay for node
replacement for a 1TB node, but I imagine this would only be used on higher
IOPS r/w nodes with smaller densities, so 100GB would be about a minute of
delay only, already within the timeframes of an AWS node
replacement/instance restart.

[Attachment #3 (text/html)]

<div dir="ltr">Does anyone have experience tooling written to support this \
strategy:<br><br>Use case: run cassandra on i3 instances on ephemerals but \
synchronize the sstables and commitlog files to the cheapest EBS volume type (those \
have bad IOPS but decent enough throughput)<br><br>On node replace, the startup \
script for the node, back-copies the sstables and commitlog state from the EBS to the \
ephemeral.<br><br>As can be seen:  <a \
href="https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/EBSVolumeTypes.html">https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/EBSVolumeTypes.html</a><br><br>the \
(presumably) spinning rust tops out at 2375 MB/sec (using multiple  EBS volumes \
presumably) that would incur about a ten minute delay for node replacement for a 1TB \
node, but I imagine this would only be used on higher IOPS r/w nodes with smaller \
densities, so 100GB would be about a minute of delay only, already within the \
timeframes of an AWS node replacement/instance restart.  <br><br><br></div>



[prev in list] [next in list] [prev in thread] [next in thread] 

Configure | About | News | Add a list | Sponsored by KoreLogic