As knowledge analytics use instances develop, components of scalability and concurrency change into essential for companies. Your analytic answer structure ought to have the ability to deal with giant knowledge volumes at excessive concurrency and with out compromising pace, thereby delivering a scalable high-performance analytics atmosphere.
Amazon Redshift Serverless offers a totally managed, petabyte-scale, auto scaling cloud knowledge warehouse to assist high-concurrency analytics. It presents knowledge analysts, builders, and scientists a quick, versatile analytic atmosphere to realize insights from their knowledge with optimum price-performance. Redshift Serverless auto scales throughout utilization spikes, enabling enterprises to cost-effectively assist meet altering enterprise calls for. You may profit from this simplicity with out altering your present analytics and enterprise intelligence (BI) functions.
To assist meet demanding efficiency wants like excessive concurrency, utilization spikes, and quick question response instances whereas optimizing prices, this publish proposes utilizing Redshift Serverless. The proposed answer goals to handle three key efficiency necessities:
- Help hundreds of concurrent connections with excessive availability by utilizing a number of Redshift Serverless endpoints behind a Community Load Balancer
- Accommodate a whole lot of concurrent queries with low-latency service stage agreements by means of scalable and distributed workgroups
- Allow subsecond response instances for brief queries towards giant datasets utilizing the quick question processing of Amazon Redshift
The steered structure makes use of a number of Redshift Serverless endpoints accessed by means of a single Community Load Balancer shopper endpoint. The Community Load Balancer evenly distributes incoming requests throughout workgroups. This improves efficiency and reduces latency by scaling out sources to satisfy excessive throughput and low latency calls for.
Resolution overview
The next diagram outlines a Redshift Serverless structure with a number of Amazon Redshift managed VPC endpoints behind a Community Load Balancer.

The next are the principle elements of this structure:
- Amazon Redshift knowledge sharing – This lets you securely share reside knowledge throughout Redshift clusters, workgroups, AWS accounts, and AWS Areas with out manually shifting or copying the information. Customers can see up-to-date and constant info in Amazon Redshift as quickly because it’s up to date. With Amazon Redshift knowledge sharing, the ingestion could be carried out on the producer or client endpoint, permitting the opposite client endpoints to learn and write the identical knowledge and thereby enabling horizontal scaling.
- Community Load Balancer – This serves as the one level of contact for shoppers. The load balancer distributes incoming site visitors throughout a number of targets, comparable to Redshift Serverless managed VPC endpoints. This will increase the provision, scalability, and efficiency of your software. You may add a number of listeners to your load balancer. A listener checks for connection requests from shoppers, utilizing the protocol and port that you simply configure, and forwards requests to a goal group. A goal group routes requests to a number of registered targets, comparable to Redshift Serverless managed VPC endpoints, utilizing the protocol and the port quantity that you simply specify.
- VPC – Redshift Serverless is provisioned in a VPC. By making a Redshift managed VPC endpoint, you allow non-public entry to Redshift Serverless from functions in one other VPC. This design permits you to scale by having a number of VPCs as wanted. The VPC endpoint offers a dedicate non-public IP for every Redshift Serverless workgroup for use because the goal teams on the Community Load Balancer.
Create an Amazon Redshift managed VPC endpoint
Full the next steps to create the Amazon Redshift managed VPC endpoint:
- On the Redshift Serverless console, select Workgroup configuration within the navigation pane.
- Select a workgroup from the record.
- On the Knowledge entry tab, within the Redshift managed VPC endpoints part, select Create endpoint.
- Enter the endpoint title. Create a reputation that’s significant in your group.
- The AWS account ID might be populated. That is your 12-digit account ID.
- Select a VPC the place the endpoint might be created.
- Select a subnet ID. In the most typical use case, it is a subnet the place you’ve got a shopper that you simply wish to connect with your Redshift Serverless occasion.
- Select which VPC safety teams so as to add. Every safety group acts as a digital firewall to regulate inbound and outbound site visitors to sources protected by the safety group, comparable to particular digital desktop situations.
The next screenshot reveals an instance of this workgroup. Word down the IP tackle to make use of throughout the creation of the goal group.

Repeat these steps to create all of your Redshift Serverless workgroups.
Add VPC endpoints for the goal group for the Community Load Balancer
So as to add these VPC endpoints to the goal group for the Community Load Balancer utilizing Amazon Elastic Compute Cloud (Amazon EC2), full the next steps:
- On the Amazon EC2 console, select Goal teams below Load Balancing within the navigation pane.
- Select Create goal group.
- For Select a goal sort, choose Cases to register targets by occasion ID, or choose IP addresses to register targets by IP tackle.
- For Goal group title, enter a reputation for the goal group.
- For Protocol, select TCP or TCP_UDP.
- For Port, use 5439 (Amazon Redshift port).
- For IP tackle sort, select IPv4 or IPv6. This selection is out there provided that the goal sort is Cases or IP addresses and the protocol is TCP or TLS.
- You could affiliate an IPv6 goal group with a dual-stack load balancer. All targets within the goal group will need to have the identical IP tackle sort. You may’t change the IP tackle sort of a goal group after you create it.
- For VPC, select the VPC with the targets to register.
- Depart the default picks for the Well being checks part, Attributes part, and Tags part.
Create a load balancer
After you create the goal group, you possibly can create your load balancer. We suggest utilizing port 5439 (Amazon Redshift default port) for it.
The Community Load Balancer serves as a single-access endpoint and might be used on connections to succeed in Amazon Redshift. This lets you add extra Redshift Serverless workgroups and improve the concurrency transparently.
Testing the answer
We examined this structure to run three BI experiences with the TPC-DS dataset (cloud benchmark dataset) as our knowledge. Amazon Redshift consists of this dataset free of charge once you select to load pattern knowledge (sample_data_dev database). The set up additionally offers the queries to check the setup.

Amongst all of the queries from TPC-DS benchmark, we selected the next three to make use of as our report queries. We modified the primary two report queries to make use of a CREATE TABLE AS SELECT (CTAS) question on non permanent tables as an alternative of the WITH clause to emulate choices you possibly can see on a typical BI instrument. For our testing, we additionally disabled the end result cache to ensure that Amazon Redshift would run the queries each time.
The set of queries incorporates the creation of non permanent tables, a be part of between these tables, and the cleanup. The cleanup step drops tables. This isn’t wanted as a result of they’re deleted on the finish of the session, however this goals to simulate all that the BI instrument does.
We used Apache JMETER to simulate shoppers invoking the requests. To be taught extra about easy methods to use and configure Apache JMETER with Amazon Redshift, confer with Constructing high-quality benchmark checks for Amazon Redshift utilizing Apache JMeter.
For the checks, we used the next configurations:
- Check 1 – A single 96 RPU Redshift Serverless vs. three workgroups at 32 RPU every
- Check 2 – A single 48 RPU Redshift Serverless vs. three workgroups at 16 RPU every
We examined three experiences by spawning 100 classes per report (300 whole). There have been 14 statements throughout the three experiences (4,200 whole). All classes had been triggered concurrently.
The next desk summarizes the tables used within the take a look at.
| Desk Identify | Row Rely |
| Catalog_page | 93,744 |
| Catalog_sales | 23,064,768 |
| Customer_address | 50,000 |
| Buyer | 100,000 |
| Date_dim | 73,049 |
| Merchandise | 144,000 |
| Promotion | 2,400 |
| Store_returns | 4,600,224 |
| Store_sales | 46,086,464 |
| Retailer | 96 |
| Web_returns | 1,148,208 |
| Web_sales | 11,510,144 |
| Web_site | 240 |
Some tables had been modified by ingesting extra knowledge than what the TPC-DS schema presents on Amazon Redshift. Knowledge was reinserted on the desk to extend the scale.
Check outcomes
The next desk summarizes our take a look at outcomes.
| TEST 1 | . | Time Consumed | Variety of Queries | Value | Max Scaled RPU | Efficiency |
| Single: 96 RPUs | 0:02:06 | 2,100 | $6 | 279 | Base | |
| Parallel: 3x 32 RPUs | 0:01:06 | 2,100 | $1.20 | 96 | 48.03% | |
| Parallel 1 (32 RPU) | 0:01:03 | 688 | $0.40 | 32 | 50.10% | |
| Parallel 2 (32 RPU) | 0:01:03 | 703 | $0.40 | 32 | 50.13% | |
| Parallel 3 (32 RPU) | 0:01:06 | 709 | $0.40 | 32 | 48.03% | |
| TEST 2 | . | Time Consumed | Variety of Queries | Value | Max Scaled RPU | Efficiency |
| Single: 48 RPUs | 0:01:55 | 2,100 | $3.30 | 168 | Base | |
| Parallel: 3x 16 RPUs | 0:01:47 | 2,100 | $1.90 | 96 | 6.77% | |
| Parallel 1 (16 RPU) | 0:01:47 | 712 | $0.70 | 36 | 6.77% | |
| Parallel 2 (16 RPU) | 0:01:44 | 696 | $0.50 | 25 | 9.13% | |
| Parallel 3 (16 RPU) | 0:01:46 | 692 | $0.70 | 35 | 7.79% |
The previous desk reveals that the parallel setup was quicker than the one at a decrease price. Additionally, in our checks, despite the fact that Check 1 had double the capability of Check 2 for the parallel setup, the associated fee was nonetheless 36% decrease and the pace was 39% quicker. Primarily based on these outcomes, we will conclude that for workloads which have excessive throughput (I/O), low latency, and excessive concurrency necessities, this structure is cost-efficient and performant. Seek advice from the AWS Pricing Value Calculator for Community Load Balancer and VPC endpoints pricing.
Redshift Serverless mechanically scales the capability to ship optimum efficiency in periods of peak workloads together with spikes in concurrency of the workload. That is evident from the utmost scaled RPU ends in the previous desk.
Lately launched options of Redshift Serverless comparable to MaxRPU and AI-driven scaling weren’t used for this take a look at. These new options can improve the price-performance of the workload even additional.
We suggest enabling cross-zone load balancing on the Community Load Balancer as a result of it distributes requests from shoppers to registered targets. Enabling cross-zone load balancing will assist stability the requests among the many Redshift Serverless managed VPC endpoints no matter the Availability Zone they’re configured in. Additionally, if the Community Load Balancer receives site visitors from just one server (identical IP), you must at all times use an odd variety of Redshift Serverless managed VPC endpoints behind the Community Load Balancer.
Conclusion
On this publish, we mentioned a scalable structure that will increase the throughput of Redshift Serverless in low latency, excessive concurrency eventualities. Having a number of Redshift Serverless workgroups behind a Community Load Balancer can ship a horizontally scalable answer at the perfect price-performance.
Moreover, Redshift Serverless makes use of AI methods (at the moment in preview) to scale mechanically with workload adjustments throughout all key dimensions—comparable to knowledge quantity adjustments, concurrent customers, and question complexity—to satisfy and keep your price-performance targets.
We hope this publish offers you with beneficial steerage. We welcome any ideas or questions within the feedback part.
In regards to the Authors
Ricardo Serafim is a Senior Analytics Specialist Options Architect at AWS.
Harshida Patel is a Analytics Specialist Principal Options Architect, with AWS.
Urvish Shah is a Senior Database Engineer at Amazon Redshift. He has greater than a decade of expertise engaged on databases, knowledge warehousing and in analytics area. Exterior of labor, he enjoys cooking, travelling and spending time together with his daughter.
Amol Gaikaiwari is a Sr. Redshift Specialist centered on serving to clients notice their enterprise outcomes with optimum Redshift price-performance. He likes to simplify knowledge pipelines and improve capabilities by means of adoption of newest Redshift options.
