Amazon Redshift, launched in 2013, has undergone vital evolution since its inception, permitting clients to broaden the horizons of information warehousing and SQL analytics. In the present day, Amazon Redshift is utilized by clients throughout all industries for quite a lot of use circumstances, together with information warehouse migration and modernization, close to real-time analytics, self-service analytics, information lake analytics, machine studying (ML), and information monetization.
Amazon Redshift made vital strides in 2024, rolling out over 100 options and enhancements. These enhancements enhanced price-performance, enabled information lakehouse architectures by blurring the boundaries between information lakes and information warehouses, simplified ingestion and accelerated close to real-time analytics, and integrated generative AI capabilities to construct pure language-based purposes and increase consumer productiveness.

Figure1: Abstract of the options and enhancements in 2024
Let’s stroll by way of a few of the current key launches, together with the brand new bulletins at AWS re:Invent 2024.
Trade-leading price-performance
Amazon Redshift presents as much as thrice higher price-performance than various cloud information warehouses. Amazon Redshift scales linearly with the variety of customers and quantity of information, making it a great resolution for each rising companies and enterprises. For instance, dashboarding purposes are a quite common use case in Redshift buyer environments the place there’s excessive concurrency and queries require fast, low-latency responses. In these eventualities, Amazon Redshift presents as much as seven instances higher throughput per greenback than various cloud information warehouses, demonstrating its distinctive worth and predictable prices.
Efficiency enhancements
Over the previous few months, we’ve launched quite a lot of efficiency enhancements to Redshift. First question response instances for dashboard queries have considerably improved by optimizing code execution and lowering compilation overhead. We’ve enhanced information sharing efficiency with improved metadata dealing with, leading to information sharing first question execution that’s as much as 4 instances quicker when the info sharing producer’s information is being up to date. We’ve enhanced autonomics algorithms to generate and implement smarter and faster optimum information structure suggestions for distribution and type keys, additional optimizing efficiency. We’ve launched new RA3.massive cases, a brand new smaller dimension RA3 node sort, to supply higher flexibility in price-performance and supply an economical migration possibility for purchasers utilizing DC2.massive cases. Moreover, we’ve rolled out AWS Graviton in Serverless, providing as much as 30% higher price-performance, and expanded concurrency scaling to help extra sorts of write queries, enabling a fair higher potential to keep up constant efficiency at scale. These enhancements collectively reinforce Amazon Redshift’s focus as a number one cloud information warehouse resolution, providing unparalleled efficiency and worth to clients.
Common availability of multi-data warehouse writes
Amazon Redshift lets you seamlessly scale with multi-cluster deployments. With the introduction of RA3 nodes with managed storage in 2019, clients obtained flexibility to scale and pay for compute and storage independently. Redshift information sharing, launched in 2020, enabled seamless cross-account and cross-Area information collaboration and stay entry with out bodily shifting the info, whereas sustaining transactional consistency. This allowed clients to scale learn analytics workloads and provided isolation to assist preserve SLAs for business-critical purposes. At re:Invent 2024, we introduced the overall availability of multi-data warehouse writes by way of information sharing for Amazon Redshift RA3 nodes and Serverless. Now you can begin writing to shared Redshift databases from a number of Redshift information warehouses in only a few clicks. The written information is on the market to all the info warehouses as quickly because it’s dedicated. This enables your groups to flexibly scale write workloads similar to extract, rework, and cargo (ETL) and information processing by including compute sources of various sorts and sizes based mostly on particular person workloads’ price-performance necessities, in addition to securely collaborate with different groups on stay information to be used circumstances similar to buyer 360.
Common availability of AI-driven scaling and optimizations
The launch of Amazon Redshift Serverless in 2021 marked a big shift, eliminating the necessity for cluster administration whereas paying for what you utilize. Redshift Serverless and information sharing enabled clients to simply implement distributed multi-cluster architectures for scaling analytics workloads. In 2024, we launched Serverless in 10 extra areas, improved performance, and added help for a capability configuration of 1024 RPUs, permitting you to deliver bigger workloads onto Redshift. Redshift Serverless can also be now much more clever and dynamic with the brand new AI-driven scaling and optimization capabilities. As a buyer, you select whether or not you wish to optimize your workloads for price, efficiency, or hold it balanced, and that’s it. Redshift Serverless works behind the scenes to scale the compute up and down and deploys optimizations to satisfy and preserve the efficiency ranges, even when workload calls for change. In inner assessments, AI-driven scaling and optimizations showcased as much as 10 instances price-performance enhancements for variable workloads.
Seamless Lakehouse architectures
Lakehouse brings collectively flexibility and openness of information lakes with the efficiency and transactional capabilities of information warehouses. Lakehouse lets you use most popular analytics engines and AI fashions of your selection with constant governance throughout all of your information. At re:Invent 2024, we unveiled the subsequent era of Amazon SageMaker, a unified platform for information, analytics, and AI. This launch brings collectively broadly adopted AWS ML and analytics capabilities, offering an built-in expertise for analytics and AI with a re-imagined lakehouse and built-in governance.
Common availability of Amazon SageMaker Lakehouse
Amazon SageMaker Lakehouse unifies your information throughout Amazon S3 information lakes and Redshift information warehouses, enabling you to construct highly effective analytics and AI/ML purposes on a single copy of information. SageMaker Lakehouse supplies the pliability to entry and question your information utilizing Apache Iceberg open requirements with the intention to use your most popular AWS, open supply, or third-party Iceberg-compatible engines and instruments. SageMaker Lakehouse presents built-in entry controls and fine-grained permissions which are persistently utilized throughout all analytics engines and AI fashions and instruments. Current Redshift information warehouses may be made accessible by way of SageMaker Lakehouse in only a easy publish step, opening up all of your information warehouse information with Iceberg REST API. You can too create new information lake tables utilizing Redshift Managed Storage (RMS) as a local storage possibility. Take a look at the Amazon SageMaker Lakehouse: Speed up analytics & AI introduced at re:Invent 2024.
Preview of Amazon SageMaker Unified Studio
Amazon SageMaker Unified Studio is an built-in information and AI growth surroundings that allows collaboration and helps groups construct information merchandise quicker. SageMaker Unified Studio brings collectively performance and instruments from a mixture of standalone studios, question editors, and visible instruments accessible at this time in Amazon EMR, AWS Glue, Amazon Redshift, Amazon Bedrock, and the prevailing Amazon SageMaker Studio, into one unified expertise. With SageMaker Unified Studio, numerous customers similar to builders, analysts, information scientists, and enterprise stakeholders can seamlessly work collectively, share sources, carry out analytics, and construct and iterate on fashions, fostering a streamlined and environment friendly analytics and AI journey.
Amazon Redshift SQL analytics on Amazon S3 Tables
At re:Invent 2024, Amazon S3 launched Amazon S3 Tables, a brand new bucket sort that’s purpose-built to retailer tabular information at scale with built-in Iceberg help. With desk buckets, you may rapidly create tables and arrange table-level permissions to handle entry to your information lake. Amazon Redshift launched help for querying Iceberg information in information lakes final yr, and now this functionality is prolonged to seamlessly querying S3 Tables. S3 Tables clients create are additionally accessible as a part of the Lakehouse for consumption by different AWS and third-party engines.
Information lake question efficiency
Amazon Redshift presents high-performance SQL capabilities on SageMaker Lakehouse, whether or not the info is in different Redshift warehouses or in open codecs. We enhanced help for querying Apache Iceberg information and improved the efficiency of querying Iceberg as much as threefold year-over-year. A variety of optimizations contribute to those speed-ups in efficiency, together with integration with AWS Glue Information Catalog statistics, improved information and metadata filtering, dynamic partition elimination, quicker/parallel processing of Iceberg manifest recordsdata, and scanner enhancements. As well as, Amazon Redshift now helps incremental refresh help for materialized views on information lake tables to get rid of the necessity for recomputing the materialized view when new information arrives, simplifying the way you construct interactive purposes on S3 information lakes.
Simplified ingestion and close to real-time analytics
On this part, we share the enhancements concerning simplified ingestion and close to real-time analytics that allow you to get quicker insights over brisker information.
Zero-ETL integration with AWS databases and third-party enterprise purposes
Amazon Redshift first launched zero-ETL integration between Amazon Aurora MySQL-Suitable Version, enabling close to real-time analytics on petabytes of transactional information from Aurora. This functionality has since expanded to help Amazon Aurora PostgreSQL-Suitable Version, Amazon Relational Database Service (Amazon RDS) for MySQL, and Amazon DynamoDB, and contains extra options similar to information filtering to selectively extract tables and schemas utilizing common expressions, help for incremental and auto-refresh materialized views on replicated information, and configurable change information seize (CDC) refresh charges.
Constructing on this innovation, at re:Invent 2024, we launched help for zero-ETL integration with eight enterprise purposes, particularly Salesforce, Zendesk, ServiceNow, SAP, Fb Advertisements, Instagram Advertisements, Pardot, and Zoho CRM. With this new functionality, you may effectively extract and cargo useful information out of your buyer help, relationship administration, and Enterprise Useful resource Planning (ERP) purposes instantly into your Redshift information warehouse for evaluation. This seamless integration eliminates the necessity for advanced, customized ingestion pipelines for ingesting the info, accelerating time to insights.
Common availability of auto-copy
Auto-copy simplifies information ingestion from Amazon S3 into Amazon Redshift. This new characteristic lets you arrange steady file ingestion out of your Amazon S3 prefix and routinely load new recordsdata to tables in your Redshift information warehouse with out the necessity for added instruments or customized options.
Streaming ingestion from Confluent Managed Cloud and self-managed Apache Kafka clusters
Amazon Redshift now helps streaming ingestion from Confluent Managed Cloud and self-managed Apache Kafka clusters on Amazon EC2instances, increasing its capabilities past Amazon Kinesis Information Streams and Amazon Managed Streaming for Apache Kafka (Amazon MSK). With this replace, you may ingest information from a wider vary of streaming sources instantly into your Redshift information warehouses for close to real-time analytics use circumstances similar to fraud detection, logistics monitoring and clickstream evaluation.
Generative AI capabilities
On this part, we share the enhancements generative AI capabilities.
Amazon Q generative SQL for Amazon Redshift
We introduced the common availability of Amazon Q generative SQL for Amazon Redshift characteristic within the Redshift Question Editor. Amazon Q generative SQL boosts productiveness by permitting customers to specific queries in pure language and obtain SQL code suggestions based mostly on their intent, question patterns, and schema metadata. The conversational interface allows customers to get insights quicker with out in depth information of the database schema. It leverages generative AI to investigate consumer enter, question historical past, and customized context like desk/column descriptions and pattern queries to supply extra related and correct SQL suggestions. This characteristic accelerates the question authoring course of and reduces the time required to derive actionable information insights.
Amazon Redshift integration with Amazon Bedrock
We introduced integration of Amazon Redshift with Amazon Bedrock, enabling you to invoke massive language fashions (LLMs) from easy SQL instructions in your information in Amazon Redshift. With this new characteristic, now you can effortlessly carry out generative AI duties similar to language translation, textual content era, summarization, buyer classification, and sentiment evaluation in your Redshift information utilizing standard basis fashions (FMs) like Anthropic’s Claude, Amazon Titan, Meta’s Llama 2, and Mistral AI. You’ll be able to invoke these fashions utilizing acquainted SQL instructions, making it less complicated than ever to combine generative AI capabilities into your information analytics workflows.
Amazon Redshift as a information base in Amazon Bedrock
Amazon Bedrock Data Bases now helps pure language querying to retrieve structured information out of your Redshift information warehouses. Utilizing superior pure language processing, Amazon Bedrock Data Bases can rework pure language queries into SQL queries, permitting customers to retrieve information instantly from the supply with out the necessity to transfer or preprocess the info. A retail analyst can now merely ask “What had been my high 5 promoting merchandise final month?”, and Amazon Bedrock Data Bases routinely interprets that question into SQL, runs the question towards Redshift, and returns the outcomes—and even supplies a summarized narrative response. To generate correct SQL queries, Amazon Bedrock Data Bases makes use of database schema, earlier question historical past, and different contextual data that’s offered concerning the information sources.
Launch abstract
Following is the launch abstract which supplies the announcement hyperlinks and reference blogs for the important thing bulletins.
Trade-leading price-performance:
Reference Blogs:
Seamless Lakehouse architectures:
Reference Blogs:
Simplified ingestion and close to real-time analytics:
Reference Blogs:
Generative AI:
Reference Blogs:
Conclusion
We proceed to innovate and evolve Amazon Redshift to satisfy your evolving information analytics wants. We encourage you to check out the most recent options and capabilities. Watch the Improvements in AWS analytics: Information warehousing and SQL analytics session from re:Invent 2024 for additional particulars. In the event you want any help, attain out to us. We’re glad to supply architectural and design steerage, in addition to help for proof of ideas and implementation. It’s Day 1!
Concerning the Creator
Neeraja Rentachintala is Director, Product Administration with AWS Analytics, main Amazon Redshift and Amazon SageMaker Lakehouse. Neeraja is a seasoned know-how chief, bringing over 25 years of expertise in product imaginative and prescient, technique, and management roles in information merchandise and platforms. She has delivered merchandise in analytics, databases, information integration, software integration, AI/ML, and large-scale distributed programs throughout on-premises and the cloud, serving Fortune 500 corporations as a part of ventures together with MapR (acquired by HPE), Microsoft SQL Server, Oracle, Informatica, and Expedia.com