LOGIN | Register
Cooperation
2026 阿里云 PolarDB 开发者大会
文章语言:
EN
Share
Minutes
原文
会议摘要
PolarDB, as a leading cloud-native database globally, significantly improves performance and efficiency, reduces enterprise costs through features such as storage and computing separation, and multi-mode data integration. In the AI era, PolarDB is committed to database intelligence, incorporating an AI engine to achieve lake-warehouse integration, support vector retrieval, graph queries, and help enterprises build intelligent data platforms. Through collaboration with Intel, using CXL technology to enhance performance, ensure data security and high availability. By partnering with Alibaba Cloud's Bailian platform, PolarDB provides a one-stop solution, supporting enterprise intelligent transformation and showcasing new capabilities in data processing in the AI era.
会议速览
Integration of AI and database technology: exploring the fundamental transformation of data infrastructure.
In the context of deep integration of AI large models and database technology, the conference discussed how data infrastructure can adapt to AI development, emphasized the transformation from traditional business support systems to data value engines and AI native access points, and proposed the need for fundamental paradigm shifts to address the data bottleneck issues in the implementation of AI.
AI Native Database: Deep Integration, Architectural Innovation, and Accelerated Industrial Implementation.
The development trend of native AI databases was discussed, emphasizing the importance of deep integration. The innovation of Polo DB in built-in AI large models, performance optimization, and integrated platform construction was introduced. At the same time, it was mentioned that continuous breakthroughs in infrastructure define cost-effectiveness in the AI era, as well as the progress of Polo DB in accelerating industrial landing. The importance of cooperation with customers was emphasized to jointly promote the advancement of database technology.
AI-Ready Cloud-Native Database: Integration of Hot Data and Big Models
In the development of AI, data and data processing capabilities are seen as the key drivers of super intelligence. The Senior Vice President of Alibaba Cloud Intelligence Group shared the concept of cloud-native databases ready for AI, emphasizing the embedding of AI large models into databases to solve data gravity and sovereignty issues, achieve the combination of hot data and model operator capabilities, promote the organic fusion of real-time memory and model intelligence, use high-quality real-time hot data as the fuel for intelligent engines, and drive intelligent interactions to occur at the point where data is generated in real-time.
PolarDB: From cloud-native database to AI-ready intelligent data engine
Shared key ideas for PolarDB's evolution from cloud native to AI, including Lake base architecture, unified metadata management, multimodal data management, and engine support. Emphasized the importance of hardware innovation, multimodal retrieval, model operatorization, AI town support and development, and customer trust. PolarDB has earned the trust of over 20,000 customers, ranking first in the market for six consecutive years, becoming a globally recognized leader. Looking ahead, PolarDB will serve as an intelligent data engine, advancing towards the era of super artificial intelligence together with developers and customers.
Database Development Trends and Enterprise Reshaping Strategies in the AI Era
Shared IDC's predictions on the growth of global intelligent bodies and tokens, emphasizing the importance of data management and cost control in the AI era, and how to reshape enterprises through data and AI, including productivity tools, business processes, user experience, etc., to help businesses adapt to future development and achieve the goals of earning money, saving money, controlling risks, and innovating.
Building an intelligent enterprise data architecture: transforming from the data plane to the business activity plane.
The discussion on how businesses can optimize operations and decision-making through the construction of intelligent data architecture was conducted, emphasizing the importance of data plane, control plane, synthesis plane, and business activity plane. The concept of 4D data was introduced, including distributed, dynamic, diverse, and dark data, and it was pointed out that leveraging these data is crucial for improving business efficiency. Simultaneously, a comparison was made between human-centered and AI-centered data processing methods, introducing new trends such as multimodal data, strategy automation, and autonomous insight generation, as well as knowledge discovery and intelligent agent participation on the business activity plane.
AI integration with databases: the key force driving enterprise transformation.
The conversation delved into the importance of the integration of AI and databases for the future impact on enterprises. It pointed out that AI-ready organizations experience significant growth in business efficiency and profits, and that databases play a core role in the AI era by storing facts and context to support intelligent actions. The global database market is expected to grow at an average annual rate of 12.8%, with China growing at 18.7%, with relational databases leading the way. Databases are evolving towards AI nativeness, integrating non-patterned and multimodal data, while cloud hosting and open standards are becoming trends, signaling a new path for enterprise transformation.
Analysis of the six major capabilities required for databases in the age of gentle AI
The conversation delved into the six major capabilities required by the database in the era of gentle AI, including federated real-time data access, native vector and multimodal semantic processing, integration of transactions and analytics, autonomous operation and intelligent data management, event-driven and streaming response, zero trust and autonomous data governance. It emphasized the role of federated computing in data privacy protection and sharing, the trend of vector database integration, and the importance of intelligent agents in real-time decision-making in business environments. It also pointed out that databases are transitioning from traditional storage query systems to intelligent bases, supporting real-time perception and decision-making of intelligent agents. Finally, it is recommended that industry users and developers choose appropriate databases and partners to adapt to the needs of the gentle AI era.
Cloud deployment leads the growth of the database market, and Alibaba Cloud's PolarDB has significant advantages.
Global database deployment trends show that cloud deployment is growing rapidly, while private deployment is tending to stagnate or decrease. Alibaba Cloud has taken the leading position in the market share of cloud-deployed relational database management systems with PolarDB, its advantages including ultimate performance, financial-grade high availability, zero-invasive migration capabilities, ecosystem compatibility, and support for intelligent application scenarios, helping companies innovate their businesses and win in the era of intelligent business.
Exploring the development path of PolarDB from cloud native to AI native
Shared the evolution path of PolarDB in the AI era, from cloud-native high-concurrency processing and separate storage, to AI-ready stage unified management of vector data and multimodal data, and then to AI-native self-building capabilities and unified management of universal data. Emphasized key upgrades such as limit list, CXL technology, integrated architecture and model operators, aiming to enhance the elasticity, scalability, and AI integration capabilities of the database, and help efficient operation of enterprise-level applications and intelligent body applications.
PolarDB adapts to the AI era: Innovation in integrated data lake and multi-modal data management.
PolarDB, through its lake-warehouse integrated architecture and multi-model data management, adapts to the demands of the AI era, achieving unified management of all-domain data, supporting open-format data, providing vertical data set management, incorporating AI nodes, and being compatible with the AI ecosystem. It aims to become an intelligent application development platform, unlocking the value of data.
Alibaba Cloud BaiLian: Innovating collaboratively between large models and databases, building a data application system without borders.
The sharing focuses on the Alibaba Cloud Bai Lian platform, introducing its role as a large-scale model integrated development platform, and how it achieves deep integration of large models and databases through model operatorization, data collaboration, and data non-out-of-domain system, promoting innovation in data processing and application development. It emphasizes the flexible use of models in data flows, combined with the agent platform and atomic operators, to ensure data security while completing the full process application from models to data.
Integration and value of data and large models in the era of enterprise agents.
Discussed the compatibility between internal business data and large models, emphasizing the importance of data form in model selection. Introduced various operators for data processing and their applications, such as document retrieval enhancement, image and text data processing, etc. Proposed the concept of multi-modal long-term memory and looked forward to the value growth that the combination of data and large models will bring to businesses in the AI era.
PolarDB: Exploring and Practicing the Transition from Cloud-Native to AI-Native Database
Shared the development of PolarDB in the past two years, with the number of customers doubling and business growing by 40%, adding 5000 GPUs for AI computing. Emphasized PolarDB's exploration in the field of AI, including multi-mode data processing, support for various types of searches, AI model integration, and AI-assisted database operations. Proposed four pillars of AI-native databases, aiming to achieve one-stop data processing, address challenges of massive data, and drive the transition of databases from cloud-native to AI-native.
PolarDB Lake Base与PolarSearch:多模数据管理与内置查询引擎的创新实践
Discussed the capabilities of PolarDB Lake Base in managing massive data, particularly in handling original data and AI-generated data, as well as the advantages of the built-in vector, text, and graph query engines in PolarSearch. Emphasized on applications such as data not migrating, hardware optimization utilization, and real-time public opinion monitoring, showcasing the actual implementation effects in industries such as the automotive sector.
PolarDB's AI-native exploration: built-in models, intelligent agent support, and AI assistant.
The speaker introduced the exploration of PolarDB in the field of AI native, including built-in AI models to achieve intelligent prediction, support for database features of intelligent bodies, solving memory bottleneck issues using CXL technology, and optimizing database operations through AI assistants. PolarDB is committed to the development of multi-modal search, built-in AI engine, intelligent database assistants, etc., aiming to enhance the level of intelligence and operational efficiency of the database.
Translation: Practical case sharing of Ideal Car's smart electric vehicle and database.
Since its establishment in 2015, Ideal Motors has shared how it integrates cloud architecture, utilizes public and private clouds such as Alibaba Cloud, and combines storage engines like MySQL, PG, PolarDB, and graph databases to support the data and services generated by 1.5 million smart electric cars. This showcases a deep integration of smart electric cars with databases and AI products in practice.
PolarDB helps Ideal Car to solve big data storage and AI inference challenges.
Facing the challenges of intelligent car big data storage and AI reasoning services, Ideal Car has successfully solved the problems of data storage costs, operational efficiency, and AI reasoning service stability by collaborating with PolarDB, utilizing its computing storage separation, massive database architecture, and AI capabilities. This has enabled the rapid implementation of enterprise-level AI applications, improving service stability and development efficiency.
The integration of Ideal Cars and PolarDB database with AI technology improves data processing efficiency.
Discussed the issue of low efficiency in traditional data extraction processes in the AI era, proposed to use PolarDB's Data Agent combined with small models and large models to achieve efficient conversion from natural language to SQL, with accuracy increased to 95%. At the same time, by utilizing PolarDB's AI-ready database data lake platform, efficient management and utilization of multimodal data can be achieved, as well as the successful launch of data engineering projects, including data extraction, cleaning, labeling, and retraining functions, aiming to reduce model training costs and optimize business processes.
PolarDB empowers AI emotional companionship: personalized interaction new paradigm of "thousand people, thousand faces".
The sharing focuses on the field of emotional companionship AI, discussing how to achieve the leap from simple replies to truly understanding users and responding instantly. It introduced the Puzzle X product launched by Du Xiaoman, which has achieved remarkable results as an AI native application in overseas markets, especially ranking first in emotional companionship app downloads in the Japanese market. The core of the success lies in building personalized emotional companionship for thousands of users, using emotional companionship psychological models and Foster trading mindset driving models to ensure the authenticity of character growth curves, avoiding AI that pleases too much, and creating a truly interactive experience that makes users feel that the characters are real.
Building a virtual companion who understands you and replies instantly: solving the three major challenges of memory, understanding, and response speed.
By collaborating with PolarDB, adopting a 4096-dimensional embedding model, graph vector AI model, and memory caching technology, we successfully addressed the challenges faced by virtual companions AI in memory, understanding causal relationships, and rapid responses, achieving a more intelligent and user-friendly AI companion experience. In the future, we will continue to explore ultra-personalized long-term memory, multimodal cognitive reasoning, and performance cost optimization, creating a more intelligent virtual partner together.
Intel and Alibaba Cloud jointly explore the evolution and innovation of database computing power in the AI era.
Intel shared their thoughts on the evolution of database computing power in the AI era, focusing on the advantages of the Xeon 6 processor in boosting database performance, supporting CXL memory pooling, and the achievements in promoting cloud-native database applications through collaboration with Alibaba Cloud. The Xeon 6 processor integrates multiple engines to significantly improve efficiency in data processing and AI workloads, especially in data preparation and AI model acceleration. The jointly developed CXL 2.0 switch technology between the two parties has greatly enhanced the scalability of databases, providing powerful computing support for the AI era.
Polo DB Distributed Database: Enterprise-level intelligent enhancement and cloud-native architecture innovation.
The head of Polo DB distributed database explained the core needs of enterprise-level databases, including security, stability, performance and cost optimization, high-concurrency service expansion, and intelligent innovation. Through DB Stack architecture innovation, the separation of data plane and control plane is achieved, ensuring data access security in cases of catastrophic failures. At the same time, Polo DBX is integrating multiple workloads, leveraging cloud infrastructure to achieve automatic resource allocation and elastic capabilities, serving high-demand scenarios such as the financial industry, and promoting the transformation and upgrade of database technology.
Data Security and High Availability: A Comprehensive Strategy for Enterprise Data Protection
Discussed the four stages of enterprise data security: pre-defense, mid-term measures, post-audit, and bottom-line recovery mechanisms, emphasizing the importance of fine-grained permission management, data encryption, end-to-end auditing, and multi-level rollback. Introduced the disaster recovery mode under the high availability standards of the financial industry, such as 4G disaster recovery, achieving an RPO of 0 through arbitration control clusters, with a city-wide RTO of 2.5 seconds and a remote RTO of 16 seconds for fast failure switchover. At the same time, optimized storage engines and caching systems, improved read performance, reduced storage costs through data compression and archiving mechanisms, and achieved fully automatic archiving rule configuration.
Key to Intelligent Empowerment in Enterprise Applications: Accurate Knowledge Search and Intelligent Entity Building
Discussed the challenges and solutions for implementing intelligent empowerment in enterprise-level applications, emphasizing the importance of accurate knowledge search and intelligent body construction. Through the case of PolarDB distributed database, it demonstrated how to utilize technologies such as vector index, full-text index, and expert-level experience to build intelligent bodies, achieve precise context construction and automated operational suggestions, thereby enhancing the intelligence level of enterprise-level applications.
Industrial and Commercial Bank of China's Database Transformation Practice: Challenges and Goals
The Software Development Center of Industrial and Commercial Bank of China shared its practice of large-scale database transformation based on PolarDB in the context of tense international situation. The transformation faced challenges such as service continuity, data security, large-scale cutting and flow, and operational automation, aiming to achieve a balance between business innovation and autonomous control, promote the integration of ID architecture system and comprehensive information creation, ensure a smooth and efficient transition, co-build with technology companies, accelerate the realization of autonomous control and innovation, and provide a high-level practice case of self-reliance and self-strengthening in technology for the financial industry.
Collaborate to create a high availability and secure data solution.
After more than a year of close collaboration, both parties have jointly developed and optimized multiple data management and security features, including high availability driver connections, intelligent security policies, and recycle bin size limits, aimed at enhancing data security and system stability. In addition, solutions for large-scale operation and maintenance scenarios were introduced, such as end-to-end tracking, automatic killing of large transactions, and integrated management platforms, effectively supporting efficient operation and maintenance of tens of thousands of database nodes.
Transformation of financial-grade distributed databases: solution for cutting-edge streaming and prospects for AI capabilities landing.
Key solutions for the transformation of financial-grade distributed databases were discussed, including the implementation of large-scale fast switch flow technology, the integration of forward and reverse switch flow capabilities, and the docking of automated platforms to ensure business continuity and data security. In addition, the emphasis was placed on reducing research and development costs, improving automated operation and maintenance capabilities, and cooperation directions in core application carrying capacity, security level improvement, and AI capability grounding in the future, aiming to support the high-quality transformation of the financial industry.
PolarDB helps GoTo Financial Credit efficiently migrate and optimize costs.
Introduced the widespread application and rapid growth of GoTo Financial's credit services in the Indonesian market, focusing on the practice of migrating from overseas cloud platform to Alibaba Cloud PolarDB. Significant results include a 35% performance improvement, a 50% cost reduction, and the advantages of PolarDB in terms of stability, automated operations, and cost optimization.
Multi-modal fusion engine and AI-driven risk control data processing optimization.
The sharing discussed how through building a multimodal fusion engine in the risk control decision-making scenario, data processing decentralization can be reduced, and data governance and monitoring efficiency can be improved. At the same time, with the help of AI technology, tasks such as data preparation are automated, allowing professionals to focus more on strategy adjustments and decision-making, significantly increasing work efficiency and productivity.
要点回答
Q:What are the specific initiatives and achievements of PolarDB in terms of technological innovation?
A:PolarDB continues to break through in architectural innovation, not only innovating in hardware with innovations such as three-tier decoupling, but also, with the efforts of the database team, once again topping the TPC-C list, demonstrating good cost control and the ability to offer inclusive solutions to the huge computational power demand. At the same time, through a storage-compute separation architecture and built-in vector retrieval capabilities, it provides a one-stop platform supporting data storage, transaction processing, and AI inference, simplifying developers' workflow.
Q:What are the characteristics and advantages of PolarDB in industrial landing?
A:Driven by market demand, PolarDB has successfully served core systems in industries such as finance, government, and telecommunications through its integrated architecture and storage-compute separation, as well as a one-stop platform. This has proven its ability to support both TPAP and AI workloads simultaneously, making it the preferred path for enterprise intelligence. This success is attributed to the extensive trust and cooperation of customers, collectively driving improvements in product performance, cost, and usability.
Q:What is the development path of AI models and their impact on databases?
A:The development path of AI is characterized by evolving from learning from and assisting humans to self-iterating and surpassing humans, leading to super artificial intelligence. In this process, the ability to collect and process data becomes crucial for expanding AI intelligence, with database systems playing a crucial role in storing and managing massive data and processing capabilities. Embedding AI large models into databases has become an inevitable choice to solve problems such as data movement, data sovereignty, and security, and to achieve an organic combination of hot data and model operators, thereby promoting real-time, high-quality data processing and intelligent interactions.
Q:In the context of deeply integrating AI large models with database technology, how should data infrastructure evolve to meet the requirements of AI readiness?
A:We need to undergo a fundamental paradigm shift, transforming data infrastructure from traditional business support systems to data engines that can maximize data value and serve as native entrances for AI. The core of the maturity of AI databases lies in deep integration, rather than simple feature addition. For example, PolarDB has launched a version with built-in AI large models, using model operator technology to embed them in the kernel. Developers can directly call AI to complete the entire closed loop, significantly improving performance and simplifying complex architectures.
Q:What specific things have you done under the grand idea?
A:We have done the following work: first, we have developed a multimodal database integrated storage solution for agent natives, combining hardware innovations such as CXL switch to build a unified memory pool, and upgrading polo FS to support integrated lake base. At the same time, we optimized model tuning rate and achieved the organic combination of multi-tenant elastic online inference services and transaction analysis by connecting the Hundred Chain platform and Pi.
Q:In terms of unified management of original data, what upgrades have you made?
A:We have upgraded the built-in metadata management capabilities of PolarDB, now supporting zero ETL lake synchronization service. This means that when any data source in the lake base changes, its original metadata information can be automatically synchronized in real time. Additionally, we also have the ability to handle heterogeneous data from multiple sources, and support seamless embedding, tagging, labeling, and feature extraction operations in metadata management.
Q:What progress have you made in terms of data retrieval and processing capabilities?
A:We have enhanced the multi-mode data retrieval and processing capabilities of PolarDB, and further promoted model computation, making it possible to provide efficient inference services within the database, allowing model calls and token consumption to be more contextualized, thereby improving resource utilization efficiency.
Q:How do you approach agent development and support?
A:We provide agent development and operation deployment support, especially focusing on key points such as long short-term memory, vectorized embedding, and feature extraction. We abstract these capabilities in PolarDB, supporting full-text indexing, graph indexing, and vector indexing, to help seamlessly support AI development in full-text data, and simplify the process of intelligent development through zero-code development.
Q:How is the development and customer recognition of PolarDB in the age of AI?
A:PolarDB has evolved from a cloud-native database to an AI-ready cloud-native database, with key evolutionary features such as Lake base, multi-model operatorization, A-performance support and development, and hardware innovation breakthroughs. Currently, PolarDB has won the trust of over 20,000 customers, with a deployment scale of over 3 million cores, covering 86 global availability zones, and has received a series of recognitions globally, such as maintaining its leadership position in the Gartner Cloud DBMS Magic Quadrant for six consecutive years.
Q:What are the development trends of future databases and the demand for databases in the AI era?
A:With the explosive growth of smart agents and tokens, enterprises will face challenges in managing and orchestrating smart agents, achieving synergy between human and smart agents, governance, security control, and cost control. At the same time, data volume will grow at a rate of 2% to 14% per year, with data no longer just being a byproduct but a core production factor. Enterprises need to reshape their business processes, including productivity tools, business processes, user experience, product services, ecosystem partners, decision-making patterns, employee organizations, and business models, to adapt to future development and establish intelligent data architecture to integrate information, deliver insights, and form a data-driven culture.
Q:In the AI era, how has the role and importance of databases changed?
A:In the AI era, the role of databases is becoming more crucial and important. AI models do not store facts, but databases need to store facts, especially after the emergence of intelligent agents, databases have shifted from being read systems to action systems, and their responsibilities and roles have changed. In addition, the cost of AI failure is increasing, making the responsibility of the data layer even greater. The overall trend in the database market is growth, with relational databases taking the lead and moving towards AI-native. AI for databases will become a standard capability, enhancing the level of automation in database maintenance and management.
Q:What are the development trends of the database market?
A:The development trends of the database market include the global market size increasing from over 90 billion dollars in 2024 to over 170 billion dollars in 2029, with an annual growth rate of 12.8%. The growth rate in the Chinese market is approximately 18.7%, and relational databases still dominate the majority of the overall market share. In terms of trends, databases are evolving towards AI native integration, non-relational and multi-modal databases will accelerate their rise, and cloud multi-cloud and open standards have become the main directions of evolution.
Q:What are the six major capabilities of databases in the AI era? Why is federated real-time data access capability crucial for databases?
A:The six major capability requirements are: federal real-time data access capability, vector and multimodal semantic processing capability, real-time decision-making capability integrating things and analytics, autonomous operation and intelligent database management capability, event-driven and streaming response capability, and zero trust and autonomous data governance capability. Federal real-time data access is one of the core requirements, predicting that by 2027, 80% of agent AI applications will require real-time, contextual, ubiquitous data access. Enterprises need to solve the problem of data sharing between different departments by enabling data to be called in real-time by intelligent agents while existing locally in various departments through methods such as federated computing, thereby improving business collaboration efficiency, reducing organizational friction costs, and shortening decision-making chains.
Q:What is the development path of PolarDB in the AI era?
A:The development path of PolarDB goes from cloud-native to AI-ready and then to AI-native. In the cloud-native stage, it mainly solves high-concurrency and massive data storage issues; in the AI-ready stage, it supports unified management of vector and multimodal data; finally, in the AI-native stage, it has certain self-made abilities, providing global data management and compatibility with AI application ecosystems, such as memory management and knowledge base management.
Q:Why is the ability of Limit List so important in the AI era?
A:In the AI era, users of databases have shifted from developers to being predominantly agent-based, demanding a higher level of elasticity, linear scalability, and read/write synchronization scalability for data platforms across generations. The Limit List is our attempt to explore these capabilities.
Q:What specific improvements were made in the second upgrade in cloud-native?
A:We have upgraded the three-tier decoupled architecture comprehensively using CXL technology. By replacing the RDMA interconnection method with CXL, we have reduced the latency from microseconds to hundreds of nanoseconds, and have broken through the performance bottleneck of a single machine, with remote memory 16 times larger than local memory. At the same time, CXL connects GPU, enabling PolarDB to implement KV cache capability internally, helping accelerate inference internally in the database.
Q:What architecture does PolarDB distributed version propose to meet the demands of enterprise-level applications?
A:The distributed version of PolarDB has proposed an integrated architecture, helping users smoothly migrate from monolithic centralized applications to native enterprise-level applications, and has built a high availability data solution at the financial level, such as global multi-activity and off-site backups. This solution has been successfully implemented and verified at Industrial and Commercial Bank of China.
Q:How to incorporate the abilities of large models into the database internally?
A:We innovatively proposed the concept of model operators, enabling the ability to call on large models within the database and enjoy the technological upgrades brought by AI throughout the entire data lifecycle. Currently, the model operators have been connected with Bailing.
Q:PolarDB如何适应AI时代的发展并拥抱lake base架构?
A:PolarDB chooses to embrace the lake base architecture to break through data barriers and adapt to the more generalized and in-depth data requirements of the AI era. We will strive in two directions, architecture evolution and support for open formats, to achieve unified management of global unstructured, semi-structured, and structured data, as well as to provide capabilities for managing domain data sets, universal data management, and multi-mode data retrieval.
Q:How is the built-in capability of large models based on AI nodes implemented in the integration of lake and warehouse?
A:By embedding AI nodes internally in PolarDB and providing a fully managed Ray processing framework, we have achieved data processing and inference happening inside the database, enabling the use of large models throughout the entire data processing lifecycle.
Q:What are the future development goals of PolarDB?
A:In the future, PolarDB is committed to three developments: 1) Continuously integrating the most advanced large models under the guidance of data gravity to fully explore the value of data; 2) Becoming the foundation intelligent base or intelligent platform for all intelligent integrated application developments; 3) Making the interaction of PolarDB more natural, supporting language processing and multi-modal processing, attracting more ordinary users to use it.
Q:Why is it not suitable to directly use embedding or tranquil methods to split and vectorize image-text data?
A:It is not suitable because when processing the image-text pair in the bottom right corner, using only the tranquil operator or data segmentation operator cannot solve the problem of the visual model understanding the semantic meaning of image-text.
Q:So how should we handle data containing PowerPoint presentations with charts?
A:We need to adopt a model similar to VL visual understanding model to project text vectors, text features, and visual features onto the same vector plane, forming a mixed vector pattern to identify and process these data in order to avoid information loss and improve model effectiveness.
Q:How to choose a suitable model based on different types of data in the enterprise, while considering factors such as cost, performance, efficiency, and effectiveness?
A:For each type of data form in the enterprise, dynamic combinations should be made based on different operators to build unique matching logic for the enterprise. With the database capability, various pipelines should be constructed to transform all types of data assets into data assets that are suitable for large models and have application value.
Q:What is the application and importance of long-term memory in databases?
A:Long-term memory is crucial for solving the continuity reasoning calculation problems of context and agent in complex reasoning tasks. In the process of real-time dynamic updates and processing, long-term memory can enhance the accuracy of the model, and provide personalized recommendations based on user profiles and fragments. In the future, multimodal long-term memory will play a greater role in the interaction between intelligent hardware and the real physical world.
Q:What work has PolarDB done in the field of AI exploration?
A:PolarDB has been committed to exploring the integration of software and hardware in the past year, including CCL pre-rehearsals, AMX technology applications, file storage improvements, etc., and has conducted in-depth research in the AI direction. Especially in the era of AI, facing the challenges of massive data processing, PolarDB is developing towards AI native direction, dedicated to solving four core issues: multi-mode data processing, efficient search, deep AI integration, and AI-assisted operation and maintenance.
Q:PolarDB Lake Base如何应对海量数据管理和AI场景需求?
A:PolarDB Lake Base is a product designed for the management of massive data, capable of supporting the storage and management of relational data and unstructured data (such as images). It consists of multiple systems, providing functions such as raw data management, object storage management, and data acceleration, to help users better manage and utilize massive data for training and inference work. This solution has been successfully applied in a large-scale data management scenario for a certain automotive enterprise.
Q:What query capabilities does the Polo Search product have? What are the advantages of the vector engine in Polo Search?
A:The Polo Search product integrates various query capabilities, including vector queries, graph queries, and text queries, all of which are built-in and operated within the database. The benefit of this design is that data does not need to be migrated or moved externally for queries, achieving integrated operations. The built-in vector engine can utilize a large number of optimization techniques, such as CXO memory and HNSW algorithms, to improve query efficiency. Additionally, this engine can utilize hardware facilities like GPUs to accelerate index construction, which can be completed in just tens of minutes compared to potentially hours with a CPU, greatly enhancing performance.
Q:What are the characteristics of Polo DB in terms of text engine and graph engine?
A:Polo DB's text engine supports inverted index and efficient search, similar to electric search, also integrated as a built-in engine. The graph engine, as a type of old-generation database, used to be commonly used for risk processing in the past. With the development of AI, it has become increasingly important in many scenarios, used to support applications such as long-term memory.
Q:How does Polo DB help with real-time public opinion monitoring?
A:Polo DB can query real-time data through its built-in engine, helping users to monitor and analyze public sentiment in real-time.
Q:polo DB如何实现model as Operator功能?
A:Starting from last year, Polo DB has been exploring the integration of various large and small models (such as BST models) into the database, so that customers can obtain real-time prediction results and utilize transaction and recharge data to determine user behavior trends.
Q:How does Polo DB solve the problem of short lifecycle and high concurrency of intelligent agents?
A:Polo DB has the characteristics of low-level control capability, fast creation and deletion of multi-mode agencies. It can effectively deal with the challenges brought by the short life cycle of intelligent agents and high concurrency, and support KV cache to optimize reasoning efficiency.
Q:What explorations has Polo DB made in the integration of AI and databases?
A:Polo DB makes AI native efforts in four directions, including multimodal fusion, advanced search, built-in AI engine (pod model as Operator), and AI assistant. Among them, the AI assistant achieves functions such as cluster information query and performance analysis through natural language processing, simplifying operation and maintenance tasks.
Q:Why did Ideal car choose to cooperate with PolarDB and what achievements did it achieve? How does Ideal car use PolarDB to solve the problems encountered in AI products?
A:Ideal car, due to the massive and rapidly growing data of intelligent electric vehicles, faced issues such as low development and operation efficiency of native databases, and doubled storage costs. After testing, Ideal car ultimately chose PolarDB. The architecture of separating computing and storage reduced costs by 60%, the native massive database architecture reduced operation costs by 50%, and supporting h-type capabilities increased operation efficiency three times. Ideal car solved data privacy and security issues in large-scale model inference services, as well as business response delays caused by GPU queueing, by integrating PolarDB's KV cache underlying technology and connecting to PolarDB AI inference model services. This resulted in a stability improvement of over 60%.
Q:After building the basic AI model inference service, how can we achieve rapid application landing and generate enterprise AI applications?
A:We have collaborated with PolarDB's super base to utilize its underlying PolarDB native database, high-performance storage, and AI data capabilities, integrating with the cloud to achieve 100% service stability, and improving development efficiency by over 100 times through web coding.
Q:After AI models and AI applications are implemented, what deep-level problems are faced?
A:The main issue lies in intelligent data acquisition. In the traditional process, businesses submit requirements to the data warehouse and wait for hours to days to obtain data, which is clearly unable to meet the efficient demands of the AI era.
Q:What are the main requirements in terms of intelligent data extraction?
A:The main demands include: first, business personnel are not familiar with SQL writing and need to obtain the latest and hottest data in real-time from natural language; second, support data acquisition from various relational databases (such as MySQL and PostgreSQL) using multiple protocols.
Q:How is the solution achieved for intelligent data retrieval?
A:We have integrated the data agent of PolarDB, which combines the advantages of small models and large models. The small model extracts key information, while the large model understands user semantics and performs NL to SQL optimization. After integrating enterprise internal knowledge data with a vector knowledge base, the accuracy can reach about 95%.
Q:In which aspects will the next generation of Ideal Cars cooperate with PolarDB?
A:Combine with PolarDB's AI data services and data lake capabilities to manage, query, and use massive multimodal data; at the same time, collaborate in the field of data engineering to leverage PolarDB's technical support to facilitate internal developers in training, deploying, and fine-tuning data.
Q:What was the theme shared by Lu Changqing, the person in charge of Du Xiaoman AI Entertainment Technology?
A:The theme shared by Lu Changqing is "PolarDB helps AI emotional companionship interactive new paradigm", focusing on how to upgrade AI companionship from simple replies to truly understanding users and providing instant companionship, including technical exploration and practical experience.
Q:How did Du Xiaoman's product X achieve success in overseas markets?
A:The key to the success of company X is building a customized emotional companionship for each individual, defining user profiles through the emotional companionship mental model, designing roles through Foster's transactional thinking model, and achieving genuine emotional interaction that makes users feel understood and cared for.
Q:How to solve memory problems and build an empathetic Iron Triangle solution?
A:In order to solve the memory problem, in the x adopts a 4096-dimensional embedding model and cooperates with PolarDB to store billions of dialogue data using PolarDB MySQL, as well as achieving vector retrieval at the level of hundreds of terabytes using PolarSearch. In addition, it also combines Polar Urgent Memory to solve memory confusion issues, and improves response speed through KV cache technology, achieving the effect of instant response.
Q:In the AI era, how does Intel view the evolution of database computing power?
A:Intel believes that computing power has become the core driving force behind the development of the digital economy. With the continuous high-speed growth of domestic computers and the vigorous development of large-scale model technology and native AI applications, the demand for comprehensive performance of CPUs is becoming increasingly high. For this reason, Intel launched the Xeon 6 processor last year, integrating various engines to support efficient collaboration of multiple industry workloads and meet the evolving computing power requirements.
Q:How does the Intel Xeon 6 processor perform in databases and AI fields? How does Intel collaborate with Alibaba Cloud to promote the development of cloud-native databases in the AI era?
A:The powerful Intel Xeon 6 processor provides comprehensive computing power, effectively improving database processing capabilities. Through built-in AI security, network data analysis and storage engines, as well as comprehensive support for CXL 2.0 specifications, significant performance improvements have been achieved in AI model tagging, responsive retrieval, and general database benchmarking within the database. For example, with AMX technology support, embedding performance can be improved by 256%. In addition, the Xeon 6 processor also optimizes the performance of built-in AI models in the database, such as utilizing AMX instruction sets and optimization toolkits to achieve performance improvements of 24% to 86% for WDL recommendation system models at different batch sizes. Intel and Alibaba Cloud work together to accelerate data processing and AI innovation through the comprehensive capabilities of the Xeon 6 processor. In the data preparation stage, the Xeon 6 demonstrates powerful processing capabilities and integrates a series of machine learning and artificial intelligence algorithms. By optimizing the performance of AI cloud-native databases, Intel and Alibaba Cloud are committed to the research and development of innovative technologies and solutions, such as promoting the development of open connectivity technology for CXL, and ensuring that the Xeon 6 supports the CXL memory pooling architecture solution to meet diverse AI workload requirements in databases. At the Cloud Computing Conference at the end of September 2025, both parties witnessed the release of the Polo DB database dedicated server equipped with CXL 2.0 switch technology, which significantly enhances database scalability and performance.
Q:What are the characteristics of the enterprise-level intelligent-enhanced Polo DB distributed database?
A:As an enterprise-level database product, Polo DB distributed database always adheres to security and stability as its core, realizing enterprise-level database requirements through kernel-level security design, data anti-tampering mechanisms, and data loss prevention. On this basis, it continuously improves cost-effectiveness, high concurrency service scalability, and intelligent innovation requirements. Especially in the Industrial and Commercial Bank database transformation project, Polo DB distributed database uses cloud-native technology to build a new DB stack architecture, achieving resource pooling, containerized services, and general database resource scheduling, ensuring the separation of data and control layers to provide extremely high elasticity. At the same time, Polo DBX is also moving towards integration, supporting multiple loads such as transaction loads, data analysis loads, and AI inference loads to run on the same distributed architecture, and ensuring data security and high availability through full-chain auditing and multi-level rollback mechanisms.
Q:In the 4G disaster recovery plan, why is it necessary to introduce an arbitration control cluster, and how to deal with data center-level failures?
A:In the 4G disaster recovery scenario, due to conflicts between the configuration of two data centers with four replicas and the Pockets consistency protocol, in order to solve the problems of large and small data centers, we introduced an arbitration control cluster. When a data center-level failure is detected, the control cluster will perform replica quantity downgrade operations to ensure RPU is 0. In this way, in large-scale data disaster recovery switchover scenarios, the same-city RTO can reach 2.5 seconds, and the average RTO for remote locations is 16 seconds.
Q:How to achieve unified processing and optimization for different types of workloads in PolarDB?
A:PolarDB achieves optimization of all types of workloads through a unified storage engine and ultra-low-cost in-memory engine. It has significantly improved the read performance of the caching system, achieving a 12-fold increase, and achieved sub-second real-time data transformation and synchronization latency. At the same time, to address the issue of rising storage costs, it has implemented maximum cost savings through online data compression and historical data archiving mechanisms. The archiving process is fully automatic, supporting partitioning and row-level TTL configuration.
Q:How to achieve intelligent empowerment in enterprise applications, especially ensuring the accuracy of the execution process?
A:For enterprise-level application intelligence empowerment, the key lies in providing deterministic and auditable operational processes. Currently, most agents operate in reactive mode, which may result in accumulated errors during execution. Databases in enterprise-level applications should provide precise factual data and accurate knowledge retrieval to meet the needs of the intelligent agent construction process. PolarDB's distributed computing storage layer introduces technologies such as vector indexing and full-text indexing, and enhances analytical acceleration and mixed retrieval capabilities in transaction systems to provide precise contextual information.
Q:What is the specific role of PolarDB in the process of constructing intelligent bodies?
A:In the intelligent body construction process, PolarDB plays a role in building precise context, ensuring accurate knowledge search and ML training automation planning and deep search through the use of long short-term memory mechanisms. It integrates large-scale expert-level experience in the cloud to form an operational intelligent body, with functions such as partition index recommendation, distributed deadlock detection, availability fault diagnosis, SQL performance insight, etc. It can build knowledge graphs based on expert experience, continuously reflect and iterate, and ultimately call on large models to provide analysis or operational recommendations.
Q:How does PolarDB x agent achieve intelligent question answering and operational optimization?
A:PolarDB x agent uses rag to implement intelligent question answering, can generate partitioning table structures and recommend the best indexes, and supports context-aware table structure evolution. For example, it can be easily transformed into a TTL table partitioned by day, and can provide advice on advanced functions such as enabling column store engines and suitable scenarios. In addition, it also has intelligent diagnostic capabilities, can locate and analyze the root cause based on error reporting, or call diagnostic tools to analyze the root cause of high availability switches.
Q:What are the main challenges faced by Industrial and Commercial Bank of China in the process of database transformation? What strategies has the Industrial and Commercial Bank of China adopted to achieve the goal of database transformation?
A:Industrial and Commercial Bank of China (ICBC) faces major challenges in database transformation, including ensuring service continuity, data security (such as preventing financial losses), large-scale complex transformation, building operation and maintenance automation platforms, controlling learning costs, implementing rapid switching and fallback mechanisms, and meeting strict production operation and maintenance requirements. ICBC has adopted two main strategies to achieve its database transformation goals: first, leveraging its unique advantages such as strict production operation and maintenance requirements, large-scale operational practices, and deep application design, research and development, and operation accumulation; second, collaborating with leading technology companies such as Alibaba Cloud to focus on core technology breakthroughs, accelerate the realization of independent controllability and innovation breakthroughs.
Q:In the process of collaborating with Alibaba Cloud, what improvements has Industrial and Commercial Bank made in terms of high availability, security, and emergency recovery?
A:Industrial and Commercial Bank of China has partnered with Aliyun to achieve driver-level high availability connection in terms of high availability, reducing the intermediate load balancing link; in terms of security, it has adjusted the password error lockout policy to only lock out the user who initiated the request, avoiding comprehensive impact, and increased the upper limit restriction on the recycle bin function to ensure rapid data recovery while preventing space constraints. In addition, they have jointly created a homemade operation and maintenance platform and DB stack center solution, improving full chain tracking, AWR performance report analysis capabilities, and automatic kill function for large transactions in MySQL.
Q:In large-scale rapid cutting flow solutions, what measures have you mainly taken to achieve automation and standardization?
A:By integrating the product with the industry's automation platform, we have achieved processes such as data migration, conversion, and cutting flow, and integrated the cutting flow of the business system itself. This manages the cutting flow of all business systems as a whole, enhancing automation and standardization capabilities. Currently, it is possible to achieve concurrent cutting flow of 300 nodes within 20 minutes in control.
Q:How do you achieve data replication from the new database to the old one in reverse synchronization, and ensure security?
A:We established a data replication link from the new database to the old database through reverse copying, and integrated with the industry automation platform, achieving the management of 300 shards on a single control panel, and completing the switching and rollback effects in 20 minutes.
Q:In terms of the development tool system, which directions do you mainly focus on to reduce costs and enhance operational capability?
A:Our tool system mainly focuses on two aspects: first, reducing research and development costs, such as optimizing DDL table structure, SQL writing and testing, etc.; second, enhancing automatic operation and maintenance capability, ensuring operation and maintenance work can handle a large number of batch tasks.
Q:What plans do you have for the future development direction?
A:In the future, we will continue to enhance the carrying capacity of core applications, such as decoupling of layers, isolation of exceptions, elastic scaling, etc., and strengthen security construction, including storage encryption, desensitization, and vulnerability protection. At the same time, we hope to further promote the landing application of AI technology in the field of financial-grade distributed databases through cooperation, in order to achieve proactive prediction and resolution of potential issues, and help drive the high-quality transformation and development of the financial industry.
Q:Can you please share a practical case study of using PolarDB in the GoTo Group?
A:At Go To Group, PolarDB is used in credit products, covering a variety of internal and external scenarios within the group. With the growth of user base, the business volume is increasing at a double-digit or even triple-digit rate each year. During the migration process from overseas cloud service providers to Alibaba Cloud using PolarDB, there was a significant improvement in performance and stability, while reducing costs by around 50%. The features of PolarDB, such as storage and computation separation, automatic scaling of read nodes, greatly simplify database management and operation, allowing the team to focus more on business innovation and value creation.
ALIBABA
follow
Follow
play
普通话
普通话
进入会议
1.0
0.5
0.75
1.0
1.5
2.0