If you use Amazon OpenSearch Service for time-bound knowledge like server logs, service logs, software logs, clickstreams, or occasion streams, storage value is likely one of the major drivers for the general value of your answer. During the last 12 months, OpenSearch Service has launched options which have opened up new prospects for storing your log knowledge in varied tiers, enabling you to commerce off knowledge latency, sturdiness, and availability. In October 2023, OpenSearch Service introduced assist for im4gn knowledge nodes, with NVMe SSD storage of as much as 30 TB. In November 2023, OpenSearch Service launched or1, the OpenSearch-optimized occasion household, which delivers as much as 30% price-performance enchancment over present cases in inner benchmarks and makes use of Amazon Easy Storage Service (Amazon S3) to offer 11 nines of sturdiness. Lastly, in Might 2024, OpenSearch Service introduced basic availability for Amazon OpenSearch Service zero-ETL integration with Amazon S3. These new options be a part of OpenSearch’s present UltraWarm cases, which give an as much as 90% discount in storage value per GB, and UltraWarm’s chilly storage possibility, which helps you to detach UltraWarm indexes and durably retailer not often accessed knowledge in Amazon S3.
This submit works by means of an instance that can assist you perceive the trade-offs accessible in value, latency, throughput, knowledge sturdiness and availability, retention, and knowledge entry, so to select the best deployment to maximise the worth of your knowledge and decrease the price.
Look at your necessities
When designing your logging answer, you want a transparent definition of your necessities as a prerequisite to creating good trade-offs. Rigorously look at your necessities for latency, sturdiness, availability, and price. Moreover, take into account which knowledge you select to ship to OpenSearch Service, how lengthy you keep knowledge, and the way you propose to entry that knowledge.
For the needs of this dialogue, we divide OpenSearch occasion storage into two courses: ephemeral backed storage and Amazon S3 backed storage. The ephemeral backed storage class consists of OpenSearch nodes that use Nonvolatile Reminiscence Categorical SSDs (NVMe SSDs) and Amazon Elastic Block Retailer (Amazon EBS) volumes. The Amazon S3 backed storage class consists of UltraWarm nodes, UltraWarm chilly storage, or1 cases, and Amazon S3 storage you entry with the service’s zero-ETL with Amazon S3. When designing your logging answer, take into account the next:
- Latency – if you happen to want leads to milliseconds, then you should use ephemeral backed storage. If seconds or minutes are acceptable, you may decrease your value through the use of Amazon S3 backed storage.
- Throughput – As a basic rule, ephemeral backed storage cases will present greater throughput. Situations which have NVMe SSDs, just like the im4gn, usually present one of the best throughput, with EBS volumes offering good throughput. or1 cases make the most of Amazon EBS storage for major shards whereas utilizing Amazon S3 with phase replication to cut back the compute value of replication, thereby providing indexing throughput that may match and even exceed NVMe-based cases.
- Knowledge sturdiness – Knowledge saved within the sizzling tier (you deploy these as knowledge nodes) has the bottom latency, and in addition the bottom sturdiness. OpenSearch Service gives automated restoration of knowledge within the sizzling tier by means of replicas, which give sturdiness with added value. Knowledge that OpenSearch shops in Amazon S3 (UltraWarm, UltraWarm chilly storage, zero-ETL with Amazon S3, and or1 cases) will get the advantage of 11 nines of sturdiness from Amazon S3.
- Knowledge availability – Greatest practices dictate that you just use replicas for knowledge in ephemeral backed storage. When you’ve at the very least one duplicate, you may proceed to entry your entire knowledge, even throughout a node failure. Nevertheless, every duplicate provides a a number of of value. For those who can tolerate momentary unavailability, you may cut back replicas by means of or1 cases, with Amazon S3 backed storage.
- Retention – Knowledge in all storage tiers incurs value. The longer you keep knowledge for evaluation, the extra cumulative value you incur for every GB of that knowledge. Establish the utmost period of time you should retain knowledge earlier than it loses all worth. In some circumstances, compliance necessities could prohibit your retention window.
- Knowledge entry – Amazon S3 backed storage cases usually have a a lot greater storage to compute ratio, offering value financial savings however with inadequate compute for high-volume workloads. When you have excessive question quantity or your queries span a big quantity of knowledge, ephemeral backed storage is the best selection. Direct question (Amazon S3 backed storage) is ideal for big quantity queries for occasionally queried knowledge.
As you take into account your necessities alongside these dimensions, your solutions will information your selections for implementation. That will help you make trade-offs, we work by means of an prolonged instance within the following sections.
OpenSearch Service value mannequin
To grasp easy methods to value an OpenSearch Service deployment, you want to perceive the price dimensions. OpenSearch Service has two completely different deployment choices: managed clusters and serverless. This submit considers managed clusters solely, as a result of Amazon OpenSearch Serverless already tiers knowledge and manages storage for you. If you use managed clusters, you configure knowledge nodes, UltraWarm nodes, and cluster supervisor nodes, deciding on Amazon Elastic Compute Cloud (Amazon EC2) occasion sorts for every of those features. OpenSearch Service deploys and manages these nodes for you, offering OpenSearch and OpenSearch Dashboards by means of a REST endpoint. You may select Amazon EBS backed cases or cases with NVMe SSD drives. OpenSearch Service costs an hourly value for the cases in your managed cluster. For those who select Amazon EBS backed cases, the service will cost you for the storage provisioned, and any provisioned IOPs you configure. For those who select or1 nodes, UltraWarm nodes, or UltraWarm chilly storage, OpenSearch Service costs for the Amazon S3 storage consumed. Lastly, the service costs for knowledge transferred out.
Instance use case
We use an instance use case to look at the trade-offs in value and efficiency. The associated fee and sizing of this instance are based mostly on greatest practices, and are directional in nature. Though you may count on to see comparable financial savings, all workloads are distinctive and your precise prices could range considerably from what we current on this submit.
For our use case, Fizzywig, a fictitious firm, is a big mushy drink producer. They’ve many crops for producing their drinks, with copious logging from their manufacturing line. They began out small, with an all-hot deployment and producing 10 GB of logs each day. Immediately, that has grown to three TB of log knowledge each day, and administration is mandating a discount in value. Fizzywig makes use of their log knowledge for occasion debugging and evaluation, in addition to historic evaluation over one 12 months of log knowledge. Let’s compute the price of storing and utilizing that knowledge in OpenSearch Service.
Ephemeral backed storage deployments
Fizzywig’s present deployment is 189 r6g.12xlarge.search knowledge nodes (no UltraWarm tier), with ephemeral backed storage. If you index knowledge in OpenSearch Service, OpenSearch builds and shops index knowledge constructions which are normally about 10% bigger than the supply knowledge, and you want to go away 25% free space for storing for working overhead. Three TB of each day supply knowledge will use 4.125 TB of storage for the primary (major) copy, together with overhead. Fizzywig follows greatest practices, utilizing two duplicate copies for optimum knowledge sturdiness and availability, with the OpenSearch Service Multi-AZ with Standby possibility, growing the storage have to 12.375 TB per day. To retailer 1 12 months of knowledge, multiply by three hundred and sixty five days to get 4.5 PB of storage wanted.
To provision this a lot storage, they may additionally select im4gn.16xlarge.search cases, or or1.16.xlarge.search cases. The next desk offers the occasion counts for every of those occasion sorts, and with one, two, or three copies of the info.
. | Max Storage (GB) per Node | Major (1 Copy) | Major + Duplicate (2 Copies) | Major + 2 Replicas (3 Copies) |
im4gn.16xlarge.search | 30,000 | 52 | 104 | 156 |
or1.16xlarge.search | 36,000 | 42 | 84 | 126 |
r6g.12xlarge.search | 24,000 | 63 | 126 | 189 |
The previous desk and the next dialogue are strictly based mostly on storage wants. or1 cases and im4gn cases each present greater throughput than r6g cases, which is able to cut back value additional. The quantity of compute saved varies between 10–40% relying on the workload and the occasion kind. These financial savings don’t move straight by means of to the underside line; they require scaling and modification of the index and shard technique to totally notice them. The previous desk and subsequent calculations take the final assumption that these deployments are over-provisioned on compute, and are storage-bound. You’ll see extra financial savings for or1 and im4gn, in contrast with r6g, if you happen to needed to scale greater for compute.
The next desk represents the entire cluster prices for the three completely different occasion sorts throughout the three completely different knowledge storage sizes specified. These are based mostly on on-demand US East (N. Virginia) AWS Area prices and embrace occasion hours, Amazon S3 value for the or1 cases, and Amazon EBS storage prices for the or1 and r6g cases.
. | Major (1 Copy) | Major + Duplicate (2 Copies) | Major + 2 Replicas (3 Copies) |
im4gn.16xlarge.search | $3,977,145 | $7,954,290 | $11,931,435 |
or1.16xlarge.search | $4,691,952 | $9,354,996 | $14,018,041 |
r6g.12xlarge.search | $4,420,585 | $8,841,170 | $13,261,755 |
This desk offers you the one-copy, two-copy, and three-copy prices (together with Amazon S3 and Amazon EBS prices, the place relevant) for this 4.5 PB workload. For this submit, “one copy” refers back to the first copy of your knowledge, with the replication issue set to zero. “Two copies” features a duplicate copy of all the knowledge, and “three copies” features a major and two replicas. As you may see, every duplicate provides a a number of of value to the answer. After all, every duplicate provides availability and sturdiness to the info. With one copy (major solely), you’d lose knowledge within the case of a single node outage (with an exception for or1 cases). With one duplicate, you may lose some or all knowledge in a two-node outage. With two replicas, you might lose knowledge solely in a three-node outage.
The or1 cases are an exception to this rule. or1 cases can assist a one-copy deployment. These cases use Amazon S3 as a backing retailer, writing all index knowledge to Amazon S3, as a method of replication, and for sturdiness. As a result of all acknowledged writes are endured in Amazon S3, you may run with a single copy, however with the chance of dropping availability of your knowledge in case of a node outage. If an information node turns into unavailable, any impacted indexes will likely be unavailable (purple) in the course of the restoration window (normally 10–20 minutes). Rigorously consider whether or not you may tolerate this unavailability along with your clients in addition to your system (for instance, your ingestion pipeline buffer). If that’s the case, you may drop your value from $14 million to $4.7 million based mostly on the one-copy (major) column illustrated within the previous desk.
Reserved Situations
OpenSearch Service helps Reserved Situations (RIs), with 1-year and 3-year phrases, with no up-front value (NURI), partial up-front value (PURI), or all up-front value (AURI). All reserved occasion commitments decrease value, with 3-year, all up-front RIs offering the deepest low cost. Making use of a 3-year AURI low cost, annual prices for Fizzywig’s workload offers prices as proven within the following desk.
. | Major | Major + Duplicate | Major + 2 Replicas |
im4gn.16xlarge.search | $1,909,076 | $3,818,152 | $5,727,228 |
or1.16xlarge.search | $3,413,371 | $6,826,742 | $10,240,113 |
r6g.12xlarge.search | $3,268,074 | $6,536,148 | $9,804,222 |
RIs present a simple solution to save value, with no code or structure modifications. Adopting RIs for this workload brings the im4gn value for 3 copies all the way down to $5.7 million, and the one-copy value for or1 cases all the way down to $3.2 million.
Amazon S3 backed storage deployments
The previous deployments are helpful as a baseline and for comparability. In fact, you’d select one of many Amazon S3 backed storage choices to maintain prices manageable.
OpenSearch Service UltraWarm cases retailer all knowledge in Amazon S3, utilizing UltraWarm nodes as a sizzling cache on high of this full dataset. UltraWarm works greatest for interactive querying of knowledge in small time-bound slices, equivalent to working a number of queries in opposition to 1 day of knowledge from 6 months in the past. Consider your entry patterns rigorously and take into account whether or not UltraWarm’s cache-like conduct will serve you nicely. UltraWarm first-query latency scales with the quantity of knowledge you want to question.
When designing an OpenSearch Service area for UltraWarm, you want to resolve in your sizzling retention window and your heat retention window. Most OpenSearch Service clients use a sizzling retention window that varies between 7–14 days, with heat retention making up the remainder of the complete retention interval. For our Fizzywig state of affairs, we use 14 days sizzling retention and 351 days of UltraWarm retention. We additionally use a two-copy (major and one duplicate) deployment within the sizzling tier.
The 14-day, sizzling storage want (based mostly on a each day ingestion price of 4.125 TB) is 115.5 TB. You may deploy six cases of any of the three occasion sorts to assist this indexing and storage. UltraWarm shops a single duplicate in Amazon S3, and doesn’t want extra storage overhead, making your 351-day storage want 1.158 PiB. You may assist this with 58 UltraWarm1.giant.search cases. The next desk offers the entire value for this deployment, with 3-year AURIs for the new tier. The or1 cases’ Amazon S3 value is rolled into the S3 column.
. | Sizzling | UltraWarm | S3 | Whole |
im4gn.16xlarge.search | $220,278 | $1,361,654 | $333,590 | $1,915,523 |
or1.16xlarge.search | $337,696 | $1,361,654 | $418,136 | $2,117,487 |
r6g.12xlarge.search | $270,410 | $1,361,654 | $333,590 | $1,965,655 |
You may additional cut back the price by shifting knowledge to UltraWarm chilly storage. Chilly storage reduces value by decreasing availability of the info—to question the info, you should situation an API name to reattach the goal indexes to the UltraWarm tier. A typical sample for 1 12 months of knowledge retains 14 days sizzling, 76 days in UltraWarm, and 275 days in chilly storage. Following this sample, you employ 6 sizzling nodes and 13 UltraWarm1.giant.search nodes. The next desk illustrates the price to run Fizzywig’s 3 TB each day workload. The or1 value for Amazon S3 utilization is rolled into the UltraWarm nodes + S3 column.
. | Sizzling | UltraWarm nodes + S3 | Chilly | Whole |
im4gn.16xlarge.search | $220,278 | $377,429 | $261,360 | $859,067 |
or1.16xlarge.search | $337,696 | $461,975 | $261,360 | $1,061,031 |
r6g.12xlarge.search | $270,410 | $377,429 | $261,360 | $909,199 |
By using Amazon S3 backed storage choices, you’re capable of cut back value even additional, with a single-copy or1 deployment at $337,000, and a most of $1 million yearly with or1 cases.
OpenSearch Service zero-ETL for Amazon S3
If you use OpenSearch Service zero-ETL for Amazon S3, you retain all of your secondary and older knowledge in Amazon S3. Secondary knowledge is the higher-volume knowledge that has decrease worth for direct inspection, equivalent to VPC Movement Logs and WAF logs. For these deployments, you retain the vast majority of occasionally queried knowledge in Amazon S3, and solely the newest knowledge in your sizzling tier. In some circumstances, you pattern your secondary knowledge, conserving a proportion within the sizzling tier as nicely. Fizzywig decides that they need to have 7 days of all of their knowledge within the sizzling tier. They’ll entry the remainder with direct question (DQ).
If you use direct question, you may retailer your knowledge in JSON, Parquet, and CSV codecs. Parquet format is perfect for direct question and gives about 75% compression on the info. Fizzywig is utilizing Amazon OpenSearch Ingestion, which might write Parquet format knowledge on to Amazon S3. Their 3 TB of each day supply knowledge compresses to 750 GB of each day Parquet knowledge. OpenSearch Service maintains a pool of compute models for direct question. You might be billed hourly for these OpenSearch Compute Items (OCUs), scaling based mostly on the quantity of knowledge you entry. For this dialog, we assume that Fizzywig can have some debugging classes and run 50 queries each day over in the future price of knowledge (750 GB). The next desk summarizes the annual value to run Fizzywig’s 3 TB each day workload, 7 days sizzling, 358 days in Amazon S3.
. | Sizzling | DQ Price | OR1 S3 | Uncooked Knowledge S3 | Whole |
im4gn.16xlarge.search | $220,278 | $2,195 | $0 | $65,772 | $288,245 |
or1.16xlarge.search | $337,696 | $2,195 | $84,546 | $65,772 | $490,209 |
r6g.12xlarge.search | $270,410 | $2,195 | $0 | $65,772 | $338,377 |
That’s fairly a journey! Fizzywig’s value for logging has come down from as excessive as $14 million yearly to as little as $288,000 yearly utilizing direct question with zero-ETL from Amazon S3. That’s a financial savings of 4,800%!
Sampling and compression
On this submit, now we have checked out one knowledge footprint to allow you to give attention to knowledge dimension, and the trade-offs you can also make relying on the way you need to entry that knowledge. OpenSearch has extra options that may additional change the economics by decreasing the quantity of knowledge you retailer.
For logs workloads, you may make use of OpenSearch Ingestion sampling to cut back the dimensions of knowledge you ship to OpenSearch Service. Sampling is suitable when your knowledge as an entire has statistical traits the place a component will be consultant of the entire. For instance, if you happen to’re working an observability workload, you may typically ship as little as 10% of your knowledge to get a consultant sampling of the traces of request dealing with in your system.
You may additional make use of a compression algorithm in your workloads. OpenSearch Service not too long ago launched assist for Zstandard (zstd) compression that may carry greater compression charges and decrease decompression latencies as in comparison with the default, greatest compression.
Conclusion
With OpenSearch Service, Fizzywig was capable of steadiness value, latency, throughput, sturdiness and availability, knowledge retention, and most popular entry patterns. They have been capable of save 4,800% for his or her logging answer, and administration was thrilled.
Throughout the board, im4gn comes out with the bottom absolute greenback quantities. Nevertheless, there are a few caveats. First, or1 cases can present greater throughput, particularly for write-intensive workloads. This will imply extra financial savings by means of decreased want for compute. Moreover, with or1’s added sturdiness, you may preserve availability and sturdiness with decrease replication, and due to this fact decrease value. One other issue to think about is RAM; the r6g cases present extra RAM, which accelerates queries for decrease latency. When coupled with UltraWarm, and with completely different sizzling/heat/chilly ratios, r6g cases will also be a wonderful selection.
Do you’ve a high-volume, logging workload? Have you ever benefitted from some or all of those strategies? Tell us!
In regards to the Creator
Jon Handler is a Senior Principal Options Architect at Amazon Internet Providers based mostly in Palo Alto, CA. Jon works intently with OpenSearch and Amazon OpenSearch Service, offering assist and steerage to a broad vary of shoppers who’ve vector, search, and log analytics workloads that they need to transfer to the AWS Cloud. Previous to becoming a member of AWS, Jon’s profession as a software program developer included 4 years of coding a large-scale, ecommerce search engine. Jon holds a Bachelor’s of the Arts from the College of Pennsylvania, and a Grasp’s of Science and a PhD in Pc Science and Synthetic Intelligence from Northwestern College.