Amazon Redshift: Ten years of continuous reinvention

海外精选
海外精选的内容汇集了全球优质的亚马逊云科技相关技术内容。同时,内容中提到的“AWS” 是 “Amazon Web Services” 的缩写,在此网站不作为商标展示。
0
0
{"value":"![image.png](https://dev-media.amazoncloud.cn/66586a0e60d64392b2d48563f920500c_image.png)\n\nThe \"[Amazon Redshift re-invented](https://www.amazon.science/publications/amazon-redshift-re-invented)\" research paper will be presented at a leading database conference next month. Two of the paper's authors, Rahul Pathak (top right), vice president of analytics at Amazon Web Services, and Ippokratis Pandis (bottom right), an Amazon Web Services senior principal engineer, discuss the origins of Redshift, how the system has evolved in the past decade, and where they see the service evolving in the years ahead.\n\nNearly ten years ago, in November 2012 at the first-ever Amazon Web Services (Amazon Web Services) re:Invent, Andy Jassy, then Amazon Web Services senior vice president, [announced the preview](https://www.youtube.com/watch?v=8FJ5DBLSFe4&t=3193s) of [Amazon Redshift](https://aws.amazon.com/redshift/?nc=sn&loc=0), the first fully managed, petabyte-scale cloud data warehouse. The service represented a significant leap forward from traditional on-premises data warehousing solutions, which were expensive, inflexible, and required significant human and capital resources to operate.\n\nIn a [blog post](https://www.allthingsdistributed.com/2012/11/amazon-redshift.html) on November 28, 2012, Werner Vogels, Amazon chief technical officer, highlighted the news: “Today, we are excited to announce the limited preview of [Amazon Redshift](http://aws.amazon.com/redshift), a fast and powerful, fully managed, petabyte-scale data warehouse service in the cloud.”\n\nFurther in the post, Vogels added, “The result of our focus on performance has been dramatic. Amazon.com’s data warehouse team has been piloting Amazon Redshift and comparing it to their on-premise data warehouse for a range of representative queries against a two billion row data set. They saw speedups ranging from 10x – 150x!”\n\nThat’s why, on the day of the announcement, [Rahul Pathak](https://www.linkedin.com/in/rp0229/), then a senior product manager, and the entire Amazon Redshift team were confident the product would be popular.\n\n“But we didn’t really understand how popular,” he recalls.\n\n“At preview we asked customers to sign up and give us some indication of their data volume and workloads,” Pathak, now vice president of Relational Engines at Amazon Web Services, said. “Within about three days we realized that we had ten times more demand for Redshift than we had planned for the entire first year of the service. So we scrambled right after re:Invent to accelerate our hardware orders to ensure we had enough capacity on the ground for when the product became generally available in early 2013. If we hadn’t done that preview, we would have been caught short.”\n\nThe Redshift team has been sprinting to keep apace of customer demand ever since. Today, the service is used by tens of thousands of customers to process exabytes of data daily. In June a subset of the team will present the paper “[Amazon Redshift re-invented](https://www.amazon.science/publications/amazon-redshift-re-invented) ” at a leading international forum for database researchers, practitioners, and developers, the [ACM SIGMOD/PODS Conference](https://2022.sigmod.org/) in Philadelphia.\n\nThe [paper](https://www.linkedin.com/in/ippokratis-pandis-03b1402/) highlights four key areas where Amazon Redshift has evolved in the past decade, provides an overview of the system architecture, describes its high-performance transactional storage and compute layers, details how smart autonomics are provided, and discusses how Amazon Web Services and Redshift make it easy for customers to use the best set of services to meet their needs.\n\nAmazon Science recently connected with two of the paper’s authors, Pathak, and [Ippokratis Pandis](https://www.linkedin.com/in/ippokratis-pandis-03b1402/), an Amazon Web Services senior principal engineer, to discuss the origins of Redshift, how the system has evolved over the past decade, and where they see the service evolving in the years ahead.\n\n#### **Q. Can you provide some background on the origin story for Redshift? What were customers seeking, and how did the initial version address those needs?**\n\n**A. Rahul**: We had been meeting with customers who in the years leading up to the launch of Amazon Redshift had moved just about every workload they had to the cloud except for their data warehouse. In many cases, it was the last thing they were running on premises, and they were still dealing with all of the challenges of on-premises data warehouses. They were expensive, had punitive licensing, were hard to scale, and customers couldn’t analyze all of their data. Customers told us they wanted to run data warehousing at scale in the cloud, that they didn’t want to compromise on performance or functionality, and that it had to be cost-effective enough for them to analyze all of their data.\n\nSo, this is what we started to build, operating under the code name Cookie Monster. This was at a time when customers’ data volumes were exploding, and not just from relational databases, but from a wide variety of sources. One of our early private beta customers tried it and the results came back so fast they thought the system was broken. It was about 10 to 20 times faster than what they had been using before. Another early customer was pretty unhappy with gaps in our early functionality. When I heard about their challenges, I got in touch, understood their feedback, and incorporated it into the service before we made it generally available in February 2013. This customer soon turned into one of our biggest advocates.\n\nWhen we launched the service and announced our pricing at $1000 a terabyte per year, people just couldn’t believe we could offer a product with that much capability at such a low price point. The fact that you could provision a data warehouse in minutes instead of months also caught everyone’s attention. It was a real game-changer for this industry segment.\n\n**Ippokratis**: I was at IBM Research at the time working on database technologies there, and we recognized that providing data warehousing as a cloud service was a game changer. It was disruptive. We were working with customers’ on-premises systems where it would take us several days or weeks to resolve an issue, whereas with a cloud data warehouse like Redshift, it would take minutes. It was also apparent that the rate of innovation would accelerate in the cloud.\n\nIn the on-premises world, it was taking months if not years to get new functionality into a software release, whereas in the cloud new capabilities could be introduced in weeks, without customers having to change a single line of code in their consuming applications. The Redshift announcement was an inflection point; I got really interested in the cloud, and cloud data warehouses, and eventually joined Amazon [Ippokratis joined the Redshift team as a principal engineer in Oct. 2015].\n\n#### **Q. How has Amazon Redshift evolved over the past decade since the launch nearly 10 years ago?**\n\n**A. Ippokratis:** As we highlight in the paper, the service has evolved at a rapid pace in response to customers’ needs. We focused on four main areas: 1) customers’ demand for high-performance execution of increasingly complex analytical queries; 2) our customers’ need to process more data and significantly increase the number of users who need to derive insights from that data; 3) customers’ need for us to make the system easier to use; and 4) our customers’ desire to integrate Redshift with other Amazon Web Services services, and the Amazon Web Services ecosystem. That’s a lot, so we’ll provide some examples across each dimension.\n\nOffering the leading price performance has been our primary focus since Rahul first began working on what would become Redshift. From the beginning, the team has focused on making core query execution latency as low as possible so customers can run more workloads, issue more jobs into the system, and run their daily analysis. To do this, Redshift generates C++ code that is highly optimized and then sends it to the distributor in the parallel database and executes this highly optimized code. This makes Redshift unique in the way it executes queries, and it has always been the core of the service.\n\nWe have never stopped innovating here to deliver our customers the best possible performance. Another thing that’s been interesting to me is that in the traditional business intelligence (BI) world, you optimize your system for very long-running jobs. But as we observe the behavior of our customers in aggregate, what’s surprising is that 90 percent of our queries among the billions we run daily in our service execute in less than one second. That’s not what people had traditionally expected from a data warehouse, and that has changed the areas of the code that we optimize.\n\n**Rahul:** As Ippokratis mentioned, the second area we focused on in the paper was customers’ need to process more data and to use that data to drive value throughout the organization. Analytics has always been super important, but eight or ten years ago it wasn’t necessarily mission critical for customers in the same way transactional databases were. That has definitely shifted. Today, core business processes rely on Redshift being highly available and performant. The biggest architectural change in the past decade in support of this goal was the introduction of Redshift Managed Storage, which allowed us to separate compute and storage, and focus a lot of innovation in each area.\n\n![image.png](https://dev-media.amazoncloud.cn/574d327321154ba6af106749fa6230d9_image.png)\n\nThe Redshift managed storage layer (RMS) is designed for a durability of 99.999999999% and 99.99% availability over a given year, across multiple availability zones. RMS manages both user data as well as transaction metadata.\n\nAnother big trend has been the desire of customers to query across and integrate disparate datasets. Redshift was the first data warehouse in the cloud to query Amazon S3 data, that was with [Redshift Spectrum](https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html) in 2017. Then we demonstrated the ability to run a query that scanned an exabyte of data in [S3](https://aws.amazon.com/s3/) as well as data in the cluster. That was a game changer.\n\nCustomers like [NASDAQ](https://aws.amazon.com/solutions/case-studies/nasdaq-data-lake/?did=cr_card&trk=cr_card) have used this extensively to query data that’s on local disk for the highest performance, but also take advantage of Redshift’s ability to integrate with the data lake and query their entire history of data with high performance. In addition to querying the data lake, integrated querying of transactional data stores like [Aurora](https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&aurora-whats-new.sort-order=desc) and [RDS](https://aws.amazon.com/free/database/?trk=c0fcea17-fb6a-4c27-ad98-192318a276ff&sc_channel=ps&sc_campaign=acquisition&sc_medium=ACQ-P|PS-GO|Brand|Desktop|SU|Database|Solution|US|EN|Text&s_kwcid=AL!4422!3!548665196298!e!!g!!amazon%20relational%20db&ef_id=Cj0KCQjwjN-SBhCkARIsACsrBz6AoeR9ak9rsboBVWrFbfeH8JqIrZP015Bpjd9XvElNLuTKVFDo5BUaAuF6EALw_wcB:G:s&s_kwcid=AL!4422!3!548665196298!e!!g!!amazon%20relational%20db) has been another big innovation, so customers can really have a high-performance analytics system that’s capable of transparently querying all of the data that matters to them without having to manage these complex integration processes that other systems require.\n\n![image.png](https://dev-media.amazoncloud.cn/5fd632ff39c64d20b84a0eddbf2626ea_image.png)\n\nThis diagram from the research paper illustrates how a query flows through Redshift. The sequence is described in detail on pages 2 and 3 of the paper.\n\n**Ippokratis:** The third area we focused on in the paper was ease of use. One change that stands out for me is that on-premises data warehousing required IT departments to have a DBA (data base administrator) who would be responsible for maintaining the environment. Over the past decade, the expectation from customers has evolved. Now, if you are offering data warehousing as a service, the systems must be capable of auto tuning, auto healing, and auto optimizing. This has become a big area of focus for us where we incorporate machine learning and automation into the system to make it easier to use, and to reduce the amount of involvement required of administrators.\n\n**Rahul:** In terms of ease of use, three innovations come to mind. One is [concurrency scaling](https://aws.amazon.com/redshift/features/concurrency-scaling/?nc=sn&loc=2&dn=3&refid=ea446940-00bb-4bee-9f27-d7a9a8080e4d). Similar to workload management, customers would previously have to manually tweak concurrency or reset clusters of the manually split workloads. Now, the system automatically provisions new resources and scales up and down without customers having to take any action. This is a great example of how Redshift has gotten much more dynamic and elastic.\n\nThe second ease of use innovation is [automated table optimization](https://docs.aws.amazon.com/redshift/latest/dg/t_Creating_tables.html). This is another place where the system is able to observe workloads and data layouts and automatically suggest how data should be sorted and distributed across nodes in the cluster. This is great because it’s a continuously learning system so workloads are never static in time.\n\nCustomers are always adding more datasets, and adding more users, so what was optimal yesterday might not be optimal tomorrow. Redshift observes this and modifies what's happening under the covers to balance that. This was the focus of a really interesting [graph optimization paper](https://www.amazon.science/publications/fast-and-effective-distribution-key-recommendation-for-amazon-redshift) that we wrote a few years ago about how to analyze for optimal distribution keys for how data is laid out within a multi-node parallel-processing system. We've coupled this with automated optimization and then table encoding. In an analytics system, how you compress data has a big impact because the less data you scan, the faster your queries go. Customers had to reason about this in the past. Now Redshift can automatically determine how to encode data correctly to deliver the best possible performance for the data and the workload.\n\nThe third innovation I want to highlight here is [Amazon Redshift Serverless](https://aws.amazon.com/redshift/redshift-serverless/), which we launched in public preview at re:Invent last fall. Redshift Serverless removes all of the management of instances and clusters, so customers can focus on getting to insights from data faster and not spend time managing infrastructure. With Redshift Serverless, customers can simply provision an endpoint and begin to interact with their data, and Redshift Serverless will auto scale and automatically manage the system to essentially remove all of that complexity from customers.\n\nCustomers can just focus on their data, set limits to manage their budgets, and we deliver optimal performance between those limits. This is another massive step forward in terms of ease of use because it eliminates any operations for customers. The early response to the preview has been tremendous. Thousands of customers have been excited to put Amazon Redshift Serverless through its paces over the past few months, and we’re excited about making it generally available in the near future.\n\n![image.png](https://dev-media.amazoncloud.cn/ae68110d40b54363905d6080d96a19dc_image.png)\n\nThe Amazon Redshift architecture as presented in the research paper.\n\n**Ippokratis:** A fourth area of focus in the paper is on integration with other Amazon Web Services services, and the Amazon Web Services ecosystem. Integration is another area where customer behavior has evolved from traditional BI use cases. Today, cloud data warehouses are a central hub with tight integration with a broader set of Amazon Web Services services. We provided the ability for customers to join data from the warehouse with the data lake. Then customers said they needed access to high-velocity business data in operational databases like [Aurora](https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&aurora-whats-new.sort-order=desc) and [RDS](https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&aurora-whats-new.sort-order=desc), so we provided access to these operational data stores. Then we added support for streams, as well as integration with [SageMaker](https://aws.amazon.com/sagemaker/) and [Lambda](https://aws.amazon.com/lambda/) so customers can run machine learning training and inference without moving their data, and do generic compute. As a result, we’ve converted the traditional BI system into a well-integrated set of Amazon Web Services services.\n\n**Rahul:** One big area of integration has been with our machine-learning ecosystem. With [Redshift ML](https://aws.amazon.com/redshift/features/redshift-ml/?nc=sn&loc=2&dn=6) we have enabled anyone who knows SQL to take advantage of all of our machine-learning innovation. We built the ability to create a model from the SQL prompt, which gets the data into Amazon S3 and calls Amazon SageMaker, to use automated machine learning to build the most appropriate model to provide predictions on the data.\n\nThis model is compiled efficiently and brought back into the data warehouse for customers to run very high-performance parallel inferences with no additional compute or no extra cost. The beauty of this integration is that every innovation we make within SageMaker means that Redshift ML gets better as well. This is just another means by which customers benefit from us connecting our services together.\n\nAnother big area for integration has been [data sharing](https://aws.amazon.com/redshift/features/data-sharing/?nc=sn&loc=2&dn=4). Once we separated storage and compute layers with [RA3 instances](https://aws.amazon.com/redshift/features/ra3/?nc=sn&loc=2&dn=5), we could enable data sharing, giving customers the ability to share data with clusters in the same account, and other accounts, or across regions. This allows us to separate consumers from producers of data, which enables things like modern data mesh architectures. Customers can share data without data copying, so they are transactionally consistent across accounts.\n\nFor example, users within a data-science organization can securely work from the shared data, as can users within the reporting or marketing organization. We’ve also integrated data sharing with [Amazon Web Services Data Exchange](), so now customers can search for — and subscribe to — third-party datasets that are live, up to date, and can be queried immediately in Redshift. This has been another game changer from the perspective of setting data free, enabling data monetization for third-party providers, and secure and live data access and licensing for subscribers for high-performance analytics within and across organizations. The fact that Redshift is part of an incredibly rich data ecosystem is a huge win for customers, and in keeping with customers’ desire to make data more pervasively available across the company.\n\n#### **Q. You indicate in the paper that Redshift innovation is continuing at an accelerated pace. How do you see the cloud data warehouse segment evolving – and more specifically Redshift – over the next several years?**\n\n**A. Rahul:** A few things will continue to be true as we head into the future. Customers will be generating ever more amounts of data, and they’re going to want to analyze that data more cost effectively. Data volumes are growing exponentially, but obviously customers don't want their costs growing exponentially. This requires that we continue to innovate, and find new levels of performance to ensure that the cost of processing a unit of data continues to go down.\nWe’ll continue innovating in software, in hardware, in silicon, and in using machine learning to make sure we deliver on that promise for customers. We’ve delivered on that promise for the past 10 years, and we’ll focus on making sure we deliver on that promise into the future.\n\n#### **I’m very proud of what the team has accomplished, but equally as excited about all the things we’re going to do to improve Redshift in the future.\n**\nIppokratis Pandis\n\nAlso, customers are always going to want better availability, they’re always going to want their data to be secure, and they’re always going to want more integrations with more data sources, and we intend to continue to deliver on all of those. What will stay the same is our ability to offer the-best in-segment price performance and capabilities, and the best-in-segment integration and security because they will always deliver value for customers.\n\n**Ippokratis:** It has been an incredible journey; we have been rebuilding the plane as we’ve been flying it with customers onboard, and this would not have happened without the support of Amazon Web Services leadership, but most importantly the tremendous engineers, managers, and product people who have worked on the team.\n\nAs we did in the paper, I want to recognize the contributions of Nate Binkert and Britt Johnson, who have passed, but whose words of wisdom continue to guide us. We’ve taken data warehousing, what we learned from books in school (Ippokratis earned his PhD in electrical and computer engineering from Carnegie Mellon University) and brought it to the cloud. In the process, we’ve been able to innovate, and write new pages in the book. I’m very proud of what the team has accomplished, but equally as excited about all the things we’re going to do to improve Redshift in the future.\n\nABOUT THE AUTHOR\n\n#### **Staff writer**","render":"<p><img src=\"https://dev-media.amazoncloud.cn/66586a0e60d64392b2d48563f920500c_image.png\" alt=\"image.png\" /></p>\n<p>The “<a href=\"https://www.amazon.science/publications/amazon-redshift-re-invented\" target=\"_blank\">Amazon Redshift re-invented</a>” research paper will be presented at a leading database conference next month. Two of the paper’s authors, Rahul Pathak (top right), vice president of analytics at Amazon Web Services, and Ippokratis Pandis (bottom right), an Amazon Web Services senior principal engineer, discuss the origins of Redshift, how the system has evolved in the past decade, and where they see the service evolving in the years ahead.</p>\n<p>Nearly ten years ago, in November 2012 at the first-ever Amazon Web Services (Amazon Web Services) re:Invent, Andy Jassy, then Amazon Web Services senior vice president, <a href=\"https://www.youtube.com/watch?v=8FJ5DBLSFe4&amp;t=3193s\" target=\"_blank\">announced the preview</a> of <a href=\"https://aws.amazon.com/redshift/?nc=sn&amp;loc=0\" target=\"_blank\">Amazon Redshift</a>, the first fully managed, petabyte-scale cloud data warehouse. The service represented a significant leap forward from traditional on-premises data warehousing solutions, which were expensive, inflexible, and required significant human and capital resources to operate.</p>\n<p>In a <a href=\"https://www.allthingsdistributed.com/2012/11/amazon-redshift.html\" target=\"_blank\">blog post</a> on November 28, 2012, Werner Vogels, Amazon chief technical officer, highlighted the news: “Today, we are excited to announce the limited preview of <a href=\"http://aws.amazon.com/redshift\" target=\"_blank\">Amazon Redshift</a>, a fast and powerful, fully managed, petabyte-scale data warehouse service in the cloud.”</p>\n<p>Further in the post, Vogels added, “The result of our focus on performance has been dramatic. Amazon.com’s data warehouse team has been piloting Amazon Redshift and comparing it to their on-premise data warehouse for a range of representative queries against a two billion row data set. They saw speedups ranging from 10x – 150x!”</p>\n<p>That’s why, on the day of the announcement, <a href=\"https://www.linkedin.com/in/rp0229/\" target=\"_blank\">Rahul Pathak</a>, then a senior product manager, and the entire Amazon Redshift team were confident the product would be popular.</p>\n<p>“But we didn’t really understand how popular,” he recalls.</p>\n<p>“At preview we asked customers to sign up and give us some indication of their data volume and workloads,” Pathak, now vice president of Relational Engines at Amazon Web Services, said. “Within about three days we realized that we had ten times more demand for Redshift than we had planned for the entire first year of the service. So we scrambled right after re:Invent to accelerate our hardware orders to ensure we had enough capacity on the ground for when the product became generally available in early 2013. If we hadn’t done that preview, we would have been caught short.”</p>\n<p>The Redshift team has been sprinting to keep apace of customer demand ever since. Today, the service is used by tens of thousands of customers to process exabytes of data daily. In June a subset of the team will present the paper “<a href=\"https://www.amazon.science/publications/amazon-redshift-re-invented\" target=\"_blank\">Amazon Redshift re-invented</a> ” at a leading international forum for database researchers, practitioners, and developers, the <a href=\"https://2022.sigmod.org/\" target=\"_blank\">ACM SIGMOD/PODS Conference</a> in Philadelphia.</p>\n<p>The <a href=\"https://www.linkedin.com/in/ippokratis-pandis-03b1402/\" target=\"_blank\">paper</a> highlights four key areas where Amazon Redshift has evolved in the past decade, provides an overview of the system architecture, describes its high-performance transactional storage and compute layers, details how smart autonomics are provided, and discusses how Amazon Web Services and Redshift make it easy for customers to use the best set of services to meet their needs.</p>\n<p>Amazon Science recently connected with two of the paper’s authors, Pathak, and <a href=\"https://www.linkedin.com/in/ippokratis-pandis-03b1402/\" target=\"_blank\">Ippokratis Pandis</a>, an Amazon Web Services senior principal engineer, to discuss the origins of Redshift, how the system has evolved over the past decade, and where they see the service evolving in the years ahead.</p>\n<h4><a id=\"Q_Can_you_provide_some_background_on_the_origin_story_for_Redshift_What_were_customers_seeking_and_how_did_the_initial_version_address_those_needs_22\"></a><strong>Q. Can you provide some background on the origin story for Redshift? What were customers seeking, and how did the initial version address those needs?</strong></h4>\n<p><strong>A. Rahul</strong>: We had been meeting with customers who in the years leading up to the launch of Amazon Redshift had moved just about every workload they had to the cloud except for their data warehouse. In many cases, it was the last thing they were running on premises, and they were still dealing with all of the challenges of on-premises data warehouses. They were expensive, had punitive licensing, were hard to scale, and customers couldn’t analyze all of their data. Customers told us they wanted to run data warehousing at scale in the cloud, that they didn’t want to compromise on performance or functionality, and that it had to be cost-effective enough for them to analyze all of their data.</p>\n<p>So, this is what we started to build, operating under the code name Cookie Monster. This was at a time when customers’ data volumes were exploding, and not just from relational databases, but from a wide variety of sources. One of our early private beta customers tried it and the results came back so fast they thought the system was broken. It was about 10 to 20 times faster than what they had been using before. Another early customer was pretty unhappy with gaps in our early functionality. When I heard about their challenges, I got in touch, understood their feedback, and incorporated it into the service before we made it generally available in February 2013. This customer soon turned into one of our biggest advocates.</p>\n<p>When we launched the service and announced our pricing at $1000 a terabyte per year, people just couldn’t believe we could offer a product with that much capability at such a low price point. The fact that you could provision a data warehouse in minutes instead of months also caught everyone’s attention. It was a real game-changer for this industry segment.</p>\n<p><strong>Ippokratis</strong>: I was at IBM Research at the time working on database technologies there, and we recognized that providing data warehousing as a cloud service was a game changer. It was disruptive. We were working with customers’ on-premises systems where it would take us several days or weeks to resolve an issue, whereas with a cloud data warehouse like Redshift, it would take minutes. It was also apparent that the rate of innovation would accelerate in the cloud.</p>\n<p>In the on-premises world, it was taking months if not years to get new functionality into a software release, whereas in the cloud new capabilities could be introduced in weeks, without customers having to change a single line of code in their consuming applications. The Redshift announcement was an inflection point; I got really interested in the cloud, and cloud data warehouses, and eventually joined Amazon [Ippokratis joined the Redshift team as a principal engineer in Oct. 2015].</p>\n<h4><a id=\"Q_How_has_Amazon_Redshift_evolved_over_the_past_decade_since_the_launch_nearly_10_years_ago_34\"></a><strong>Q. How has Amazon Redshift evolved over the past decade since the launch nearly 10 years ago?</strong></h4>\n<p><strong>A. Ippokratis:</strong> As we highlight in the paper, the service has evolved at a rapid pace in response to customers’ needs. We focused on four main areas: 1) customers’ demand for high-performance execution of increasingly complex analytical queries; 2) our customers’ need to process more data and significantly increase the number of users who need to derive insights from that data; 3) customers’ need for us to make the system easier to use; and 4) our customers’ desire to integrate Redshift with other Amazon Web Services services, and the Amazon Web Services ecosystem. That’s a lot, so we’ll provide some examples across each dimension.</p>\n<p>Offering the leading price performance has been our primary focus since Rahul first began working on what would become Redshift. From the beginning, the team has focused on making core query execution latency as low as possible so customers can run more workloads, issue more jobs into the system, and run their daily analysis. To do this, Redshift generates C++ code that is highly optimized and then sends it to the distributor in the parallel database and executes this highly optimized code. This makes Redshift unique in the way it executes queries, and it has always been the core of the service.</p>\n<p>We have never stopped innovating here to deliver our customers the best possible performance. Another thing that’s been interesting to me is that in the traditional business intelligence (BI) world, you optimize your system for very long-running jobs. But as we observe the behavior of our customers in aggregate, what’s surprising is that 90 percent of our queries among the billions we run daily in our service execute in less than one second. That’s not what people had traditionally expected from a data warehouse, and that has changed the areas of the code that we optimize.</p>\n<p><strong>Rahul:</strong> As Ippokratis mentioned, the second area we focused on in the paper was customers’ need to process more data and to use that data to drive value throughout the organization. Analytics has always been super important, but eight or ten years ago it wasn’t necessarily mission critical for customers in the same way transactional databases were. That has definitely shifted. Today, core business processes rely on Redshift being highly available and performant. The biggest architectural change in the past decade in support of this goal was the introduction of Redshift Managed Storage, which allowed us to separate compute and storage, and focus a lot of innovation in each area.</p>\n<p><img src=\"https://dev-media.amazoncloud.cn/574d327321154ba6af106749fa6230d9_image.png\" alt=\"image.png\" /></p>\n<p>The Redshift managed storage layer (RMS) is designed for a durability of 99.999999999% and 99.99% availability over a given year, across multiple availability zones. RMS manages both user data as well as transaction metadata.</p>\n<p>Another big trend has been the desire of customers to query across and integrate disparate datasets. Redshift was the first data warehouse in the cloud to query Amazon S3 data, that was with <a href=\"https://docs.aws.amazon.com/redshift/latest/dg/c-getting-started-using-spectrum.html\" target=\"_blank\">Redshift Spectrum</a> in 2017. Then we demonstrated the ability to run a query that scanned an exabyte of data in <a href=\"https://aws.amazon.com/s3/\" target=\"_blank\">S3</a> as well as data in the cluster. That was a game changer.</p>\n<p>Customers like <a href=\"https://aws.amazon.com/solutions/case-studies/nasdaq-data-lake/?did=cr_card&amp;trk=cr_card\" target=\"_blank\">NASDAQ</a> have used this extensively to query data that’s on local disk for the highest performance, but also take advantage of Redshift’s ability to integrate with the data lake and query their entire history of data with high performance. In addition to querying the data lake, integrated querying of transactional data stores like <a href=\"https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&amp;aurora-whats-new.sort-order=desc\" target=\"_blank\">Aurora</a> and <a href=\"https://aws.amazon.com/free/database/?trk=c0fcea17-fb6a-4c27-ad98-192318a276ff&amp;sc_channel=ps&amp;sc_campaign=acquisition&amp;sc_medium=ACQ-P%7CPS-GO%7CBrand%7CDesktop%7CSU%7CDatabase%7CSolution%7CUS%7CEN%7CText&amp;s_kwcid=AL!4422!3!548665196298!e!!g!!amazon%20relational%20db&amp;ef_id=Cj0KCQjwjN-SBhCkARIsACsrBz6AoeR9ak9rsboBVWrFbfeH8JqIrZP015Bpjd9XvElNLuTKVFDo5BUaAuF6EALw_wcB:G:s&amp;s_kwcid=AL!4422!3!548665196298!e!!g!!amazon%20relational%20db\" target=\"_blank\">RDS</a> has been another big innovation, so customers can really have a high-performance analytics system that’s capable of transparently querying all of the data that matters to them without having to manage these complex integration processes that other systems require.</p>\n<p><img src=\"https://dev-media.amazoncloud.cn/5fd632ff39c64d20b84a0eddbf2626ea_image.png\" alt=\"image.png\" /></p>\n<p>This diagram from the research paper illustrates how a query flows through Redshift. The sequence is described in detail on pages 2 and 3 of the paper.</p>\n<p><strong>Ippokratis:</strong> The third area we focused on in the paper was ease of use. One change that stands out for me is that on-premises data warehousing required IT departments to have a DBA (data base administrator) who would be responsible for maintaining the environment. Over the past decade, the expectation from customers has evolved. Now, if you are offering data warehousing as a service, the systems must be capable of auto tuning, auto healing, and auto optimizing. This has become a big area of focus for us where we incorporate machine learning and automation into the system to make it easier to use, and to reduce the amount of involvement required of administrators.</p>\n<p><strong>Rahul:</strong> In terms of ease of use, three innovations come to mind. One is <a href=\"https://aws.amazon.com/redshift/features/concurrency-scaling/?nc=sn&amp;loc=2&amp;dn=3&amp;refid=ea446940-00bb-4bee-9f27-d7a9a8080e4d\" target=\"_blank\">concurrency scaling</a>. Similar to workload management, customers would previously have to manually tweak concurrency or reset clusters of the manually split workloads. Now, the system automatically provisions new resources and scales up and down without customers having to take any action. This is a great example of how Redshift has gotten much more dynamic and elastic.</p>\n<p>The second ease of use innovation is <a href=\"https://docs.aws.amazon.com/redshift/latest/dg/t_Creating_tables.html\" target=\"_blank\">automated table optimization</a>. This is another place where the system is able to observe workloads and data layouts and automatically suggest how data should be sorted and distributed across nodes in the cluster. This is great because it’s a continuously learning system so workloads are never static in time.</p>\n<p>Customers are always adding more datasets, and adding more users, so what was optimal yesterday might not be optimal tomorrow. Redshift observes this and modifies what’s happening under the covers to balance that. This was the focus of a really interesting <a href=\"https://www.amazon.science/publications/fast-and-effective-distribution-key-recommendation-for-amazon-redshift\" target=\"_blank\">graph optimization paper</a> that we wrote a few years ago about how to analyze for optimal distribution keys for how data is laid out within a multi-node parallel-processing system. We’ve coupled this with automated optimization and then table encoding. In an analytics system, how you compress data has a big impact because the less data you scan, the faster your queries go. Customers had to reason about this in the past. Now Redshift can automatically determine how to encode data correctly to deliver the best possible performance for the data and the workload.</p>\n<p>The third innovation I want to highlight here is <a href=\"https://aws.amazon.com/redshift/redshift-serverless/\" target=\"_blank\">Amazon Redshift Serverless</a>, which we launched in public preview at re:Invent last fall. Redshift Serverless removes all of the management of instances and clusters, so customers can focus on getting to insights from data faster and not spend time managing infrastructure. With Redshift Serverless, customers can simply provision an endpoint and begin to interact with their data, and Redshift Serverless will auto scale and automatically manage the system to essentially remove all of that complexity from customers.</p>\n<p>Customers can just focus on their data, set limits to manage their budgets, and we deliver optimal performance between those limits. This is another massive step forward in terms of ease of use because it eliminates any operations for customers. The early response to the preview has been tremendous. Thousands of customers have been excited to put Amazon Redshift Serverless through its paces over the past few months, and we’re excited about making it generally available in the near future.</p>\n<p><img src=\"https://dev-media.amazoncloud.cn/ae68110d40b54363905d6080d96a19dc_image.png\" alt=\"image.png\" /></p>\n<p>The Amazon Redshift architecture as presented in the research paper.</p>\n<p><strong>Ippokratis:</strong> A fourth area of focus in the paper is on integration with other Amazon Web Services services, and the Amazon Web Services ecosystem. Integration is another area where customer behavior has evolved from traditional BI use cases. Today, cloud data warehouses are a central hub with tight integration with a broader set of Amazon Web Services services. We provided the ability for customers to join data from the warehouse with the data lake. Then customers said they needed access to high-velocity business data in operational databases like <a href=\"https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&amp;aurora-whats-new.sort-order=desc\" target=\"_blank\">Aurora</a> and <a href=\"https://aws.amazon.com/rds/aurora/?aurora-whats-new.sort-by=item.additionalFields.postDateTime&amp;aurora-whats-new.sort-order=desc\" target=\"_blank\">RDS</a>, so we provided access to these operational data stores. Then we added support for streams, as well as integration with <a href=\"https://aws.amazon.com/sagemaker/\" target=\"_blank\">SageMaker</a> and <a href=\"https://aws.amazon.com/lambda/\" target=\"_blank\">Lambda</a> so customers can run machine learning training and inference without moving their data, and do generic compute. As a result, we’ve converted the traditional BI system into a well-integrated set of Amazon Web Services services.</p>\n<p><strong>Rahul:</strong> One big area of integration has been with our machine-learning ecosystem. With <a href=\"https://aws.amazon.com/redshift/features/redshift-ml/?nc=sn&amp;loc=2&amp;dn=6\" target=\"_blank\">Redshift ML</a> we have enabled anyone who knows SQL to take advantage of all of our machine-learning innovation. We built the ability to create a model from the SQL prompt, which gets the data into Amazon S3 and calls Amazon SageMaker, to use automated machine learning to build the most appropriate model to provide predictions on the data.</p>\n<p>This model is compiled efficiently and brought back into the data warehouse for customers to run very high-performance parallel inferences with no additional compute or no extra cost. The beauty of this integration is that every innovation we make within SageMaker means that Redshift ML gets better as well. This is just another means by which customers benefit from us connecting our services together.</p>\n<p>Another big area for integration has been <a href=\"https://aws.amazon.com/redshift/features/data-sharing/?nc=sn&amp;loc=2&amp;dn=4\" target=\"_blank\">data sharing</a>. Once we separated storage and compute layers with <a href=\"https://aws.amazon.com/redshift/features/ra3/?nc=sn&amp;loc=2&amp;dn=5\" target=\"_blank\">RA3 instances</a>, we could enable data sharing, giving customers the ability to share data with clusters in the same account, and other accounts, or across regions. This allows us to separate consumers from producers of data, which enables things like modern data mesh architectures. Customers can share data without data copying, so they are transactionally consistent across accounts.</p>\n<p>For example, users within a data-science organization can securely work from the shared data, as can users within the reporting or marketing organization. We’ve also integrated data sharing with <a href=\"\" target=\"_blank\">Amazon Web Services Data Exchange</a>, so now customers can search for — and subscribe to — third-party datasets that are live, up to date, and can be queried immediately in Redshift. This has been another game changer from the perspective of setting data free, enabling data monetization for third-party providers, and secure and live data access and licensing for subscribers for high-performance analytics within and across organizations. The fact that Redshift is part of an incredibly rich data ecosystem is a huge win for customers, and in keeping with customers’ desire to make data more pervasively available across the company.</p>\n<h4><a id=\"Q_You_indicate_in_the_paper_that_Redshift_innovation_is_continuing_at_an_accelerated_pace__How_do_you_see_the_cloud_data_warehouse_segment_evolving__and_more_specifically_Redshift__over_the_next_several_years_82\"></a><strong>Q. You indicate in the paper that Redshift innovation is continuing at an accelerated pace. How do you see the cloud data warehouse segment evolving – and more specifically Redshift – over the next several years?</strong></h4>\n<p><strong>A. Rahul:</strong> A few things will continue to be true as we head into the future. Customers will be generating ever more amounts of data, and they’re going to want to analyze that data more cost effectively. Data volumes are growing exponentially, but obviously customers don’t want their costs growing exponentially. This requires that we continue to innovate, and find new levels of performance to ensure that the cost of processing a unit of data continues to go down.<br />\nWe’ll continue innovating in software, in hardware, in silicon, and in using machine learning to make sure we deliver on that promise for customers. We’ve delivered on that promise for the past 10 years, and we’ll focus on making sure we deliver on that promise into the future.</p>\n<h4><a id=\"Im_very_proud_of_what_the_team_has_accomplished_but_equally_as_excited_about_all_the_things_were_going_to_do_to_improve_Redshift_in_the_future_87\"></a>**I’m very proud of what the team has accomplished, but equally as excited about all the things we’re going to do to improve Redshift in the future.</h4>\n<p>**<br />\nIppokratis Pandis</p>\n<p>Also, customers are always going to want better availability, they’re always going to want their data to be secure, and they’re always going to want more integrations with more data sources, and we intend to continue to deliver on all of those. What will stay the same is our ability to offer the-best in-segment price performance and capabilities, and the best-in-segment integration and security because they will always deliver value for customers.</p>\n<p><strong>Ippokratis:</strong> It has been an incredible journey; we have been rebuilding the plane as we’ve been flying it with customers onboard, and this would not have happened without the support of Amazon Web Services leadership, but most importantly the tremendous engineers, managers, and product people who have worked on the team.</p>\n<p>As we did in the paper, I want to recognize the contributions of Nate Binkert and Britt Johnson, who have passed, but whose words of wisdom continue to guide us. We’ve taken data warehousing, what we learned from books in school (Ippokratis earned his PhD in electrical and computer engineering from Carnegie Mellon University) and brought it to the cloud. In the process, we’ve been able to innovate, and write new pages in the book. I’m very proud of what the team has accomplished, but equally as excited about all the things we’re going to do to improve Redshift in the future.</p>\n<p>ABOUT THE AUTHOR</p>\n<h4><a id=\"Staff_writer_99\"></a><strong>Staff writer</strong></h4>\n"}
目录
亚马逊云科技解决方案 基于行业客户应用场景及技术领域的解决方案
联系亚马逊云科技专家
亚马逊云科技解决方案
基于行业客户应用场景及技术领域的解决方案
联系专家
0
目录
关闭