Data Lake Compute is a new data architecture with closed-loop big data analysis that is lightweight, agile, easy-to-use, and cost-effective. It has a unified metadata management view that allows you to break through data silos. It combines the strengths of many cloud-based big data services to accommodate real-time and offline data analysis scenarios and comprehensively solve a wide range of data problems. Moreover, with convenient and swift data flows, it features many of the capabilities and advantages of different cloud services, making it an ideal option for enterprises setting up a data middleend.
Typical Use Cases
Data Lake Compute enables you to unify all of your different metadata views into one. In this way, you can manage and use metadata from different sources in a centralized manner, build your metadata center with agility, and switch between products and versions seamlessly. Specifically, you can easily reuse the same metadata across products.
- Agile and Versatile Data Analysis
In the big data ecosystem, Presto excels in performing interactive analysis while Spark does well in ETL tasks. Data Lake Compute provides unified syntax and lightweight clustering capabilities, so the same data can go seamlessly between engines in different scenarios. It also works with WeData so data can be imported from and exported to dozens of products and data sources, such as TencentDB and CLS. This makes the most out of the strengths of each product through convenient data flows.
Service Benefits
- Out-of-the-box: Unnecessary Ops tasks and costs are saved.
- Metadata management: Multiple data sources are supported to unify metadata management and break through data silos.
- Full coverage: Data Lake Compute comprehensively covers data analysis and application scenarios, specifically, data integration, synergy, scheduling, development, and governance.