The startup adds support for columnar storage, in-memory caching and cost-optimized AWS storage to its hybrid transactional and analytical processing platform. Credit: Thinkstock Building on its capability to concurrently run enterprise-scale transactional and analytical workloads, Splice Machine today announced the release of version 2.5 of its platform at AWS re:Invent 2016. The startup’s platform is a dual-engine relational database management system (RDBMS) powered by Apache Hadoop and Apache Spark that specializes in hybrid transactional and analytical processing (HTAP). Splice Machine uses resource isolation — separate processes and resource management for its Hadoop and Spark components — to ensure that large, complex online analytical processing (OLAP) queries don’t overwhelm time-sensitive online transaction processing (OLTP) queries. The hybrid architecture allows you to run analytical workloads and transactional workloads concurrently — a boon for use cases ranging from digital marketing to ETL acceleration, operational data lakes, data warehouse offloads, Internet of Things (IoT) applications, web, mobile and social applications and operational applications. SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe The latest release adds support for columnar storage, in-memory caching and cost-optimized storage for AWS users, among other features. At AWS re:Invent, Splice Machine demonstrated how users can leverage the new capabilities on AWS to integrate multiple compute and storage engines into an elastically scalable database that can be a relational database and data warehouse in one. “The new capabilities further emphasize the benefits of Splice Machine’s hybrid architecture,” Monte Zweben, co-founder and CEO of Splice Machine, said in a statement today. “For modern applications that need to combine fast data ingestion, web-scale transactional and analytical workloads and continuous machine learning, one storage model does not fit all. The Splice Machine SQL RDBMS tightly integrates multiple compute engines, with in-memory and persistent storage in both row-based and columnar formats. The cost-based optimizer uses new advanced statistics to find the optimal execution strategy across all these resources for OLTP and OLAP workloads.” The new capabilities of version 2.5 of the Splice Machine platform include the following: Columnar External Tables. Columnar External Tables enables hybrid columnar and row-based querying. Columnar external tables can be created in Apache Parquet, Apache ORC or text formats. Columnar Storage improves large table scans, large joins, aggregations or groupings while the native row-based storage is used for write-optimized ingestion, single-record lookups/updates and short scans. In-Memory Caching via Pinning. This feature provides the ability to move tables and columnar data files into memory for lightning-fast data access. It avoids multiple table scans or writes to high-latency file systems such as Amazon S3. Splice Machine says the capability allows data to be stored on very inexpensive storage while being very performant in-memory when required in applications. Statistics via Sketching. This feature helps solve the age-old problem that cost-based optimizers are only as good as their statistics, but most statistics are poor because statistics computation is expensive. Splice Machine utilizes the sketching library created by Yahoo! to provide very fast approximate analysis of big data statistics with bounded errors. Using sketches and histograms, Splice Machine says the cost-based optimizer can choose indexes, join orders and join algorithms with much more accuracy. Cost-Optimized Storage for AWS users. Data can be stored locally in ephemeral storage, on EBS, S3 and EFS. Depending on the workload and longevity of data, different data can be stored in different storage systems with different price/performance characteristics. Related content news analysis Skilled IT pay defined by volatility, security, and AI Foote Partners’ Q3 report on IT skills pay trends show AI and security skills were in high demand, and the value of cash-pay premiums was more volatile but their average value across a broad range of IT skills and certifications was slightly do By Peter Sayer Oct 04, 2023 6 mins Certifications Technology Industry IT Skills feature 4 reasons why gen AI projects fail Data issues are still among the chief reasons why AI projects fall short of expectations, but the advent of generative AI has added a few new twists. By Maria Korolov Oct 04, 2023 9 mins Data Science Data Science Data Science feature What a quarter century of digital transformation at PayPal looks like Currently processing a volume of payments worth over $1.3 trillion, PayPal has repeatedly staked its claim as a digital success story over the last 25 years. But insiders agree this growth needs to be constantly supported by reliable technological ar By Nuria Cordon Oct 04, 2023 7 mins Payment Systems Digital Transformation Innovation brandpost Future-Proofing Your Business with Hyperautomation By Veronica Lew Oct 03, 2023 7 mins Robotic Process Automation Podcasts Videos Resources Events SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe