Frank will share the motivation behind the 3D XPoint memory, the current shipping Optane SSD product and key values of why it is better than NAND-based SSDs, and a few use cases that exist in the Open Source space for Database usages of Optane SSDs.
Scylla Summit 2017: Migrating to Scylla From Cassandra and Others With No Dow...ScyllaDB
The session will cover the best practices to migrate existing data from Apache Cassandra to Scylla and how to do it while being online all of the time.
Scylla Summit 2017: Scylla for Mass Simultaneous Sensor Data Processing of ME...ScyllaDB
We will share Scylla adoption practices in equipment sensor data management of MES, Data Modeling Tips, Data Architecture using Scylla, configurations, and tunings.
Scylla Summit 2017: How to Use Gocql to Execute Queries and What the Driver D...ScyllaDB
This document outlines a presentation on using the GoCQL driver to execute queries against Cassandra and Scylla databases. It discusses connecting to a Cassandra cluster, executing queries, iterating over results, and using asynchronous queries. It also mentions some additional Cassandra libraries built on top of GoCQL, including gocqlx for data binding and queries, and gocassa for queries and migrations. The presentation aims to explain how GoCQL works behind the scenes and how to get started with basic querying functionality.
Scylla Summit 2017: How to Ruin Your Workload's Performance by Choosing the W...ScyllaDB
In my talk, I will present the different compaction strategies that Scylla provides, and demonstrate when it is appropriate and when it is inappropriate to use each one. I will then present a new compaction strategy that we designed as a lesson from the existing compaction strategies by picking the best features of the existing strategies while avoiding their problems.
Scylla Summit 2017: Scylla on Samsung NVMe Z-SSDsScyllaDB
I will be giving a talk about performance characterization and tuning of Scylla on Samsung NVMe SSDs. We will characterize the performance of Scylla on Samsung high-performance NVMe SSDs and show how Z-SSD ─ the Samsung ultra-low-latency NVMe drive ─ can significantly shrink the performance gap between in-memory and in-storage with Scylla.
We will further evaluate the throughput-vs-latency profile of Scylla with NVMe devices and present end-to-end latencies (from the client's viewpoint) as well as the latencies of the software/hardware stack. We will show that a Z-SSD-backed Scylla cluster can provide competitive performance to an in-memory deployment while sharply reducing costs.
Scylla Summit 2017: The Upcoming HPC EvolutionScyllaDB
In this talk, I will explain how HPC is beginning to evolve and how we use supercomputers to monitor supercomputers. First we will look at how HPC is different from cloud computing in terms of infrastructure and application architecture. Then I will discuss how those things are changing and why. Finally, I will dive into a use case of monitoring supercomputers as an application area for Scylla.
Scylla Summit 2017: A Deep Dive on Heat Weighted Load BalancingScyllaDB
This presentation discusses the "cold node problem" that occurs when a node restarts in a Cassandra cluster. When a node restarts, it loses its cached data and becomes a bottleneck. The presentation proposes a "heat weighted load balancing" solution where the cluster tracks each node's cache hit ratio and redistributes requests based on this ratio after a restart. Testing shows this solution significantly improves throughput after a node restart by distributing requests more evenly across nodes based on their "heat" or cache contents.
Kubernetes is a declarative system for automatically deploying, managing, and scaling applications and their dependencies. In this short talk, I'll demonstrate a small Scylla cluster running in Google Compute Engine via Kubernetes and our publicly-published Docker images.
ScyllaDB CTO Avi Kivity gave a keynote on how Scylla has evolved. He discussed new features in Scylla 2.0—including Materialized Views and Heat-Weighted Load Balancing, changes in monitoring—and shared our product roadmap. He also talked about our recent acquisition of Seastar.io and how it will enable us to deliver a database-as-a-service offering.
Scylla Summit 2017: Scylla's Open Source Monitoring SolutionScyllaDB
Scylla's monitoring capability has come a long way in the last year. We now have native support for Prometheus. Through scylla-grafana-monitoring, we have started providing default dashboards summarizing the most important aspects of Scylla for users. In this talk, I will cover what is currently available in our metrics, other non-standard metrics that are interesting but not available in our main dashboard, as well as our future plans for enhancement.
If You Care About Performance, Use User Defined TypesScyllaDB
Shlomi Livne, VP of R&D at ScyllaDB, presented on the performance benefits of using user-defined types (UDTs) in ScyllaDB. He explained that with traditional columns, each column has overhead and flexibility comes at a price. However, with frozen UDTs, the columns are treated as a single unit, sharing metadata and improving performance. Livne showed results of a test where UDTs with many fields outperformed traditional columns with the same number of fields. However, he noted that Scylla's row cache and Java driver performance need improvement for UDTs.
Scylla Summit 2017: How to Run Cassandra/Scylla from a MySQL DBA's Point of ViewScyllaDB
Are you a MySQL DBA or DevOps individual being asked to run Cassandra or Scylla? Feeling overwhelmed? In this talk, I will present Cassandra/Scylla operations in terms that directly relate to MySQL. I will show you comparisons between the Information Schema and the Cassandra/Scylla System keyspace(s). I will also talk about metrics available in MySQL versus Cassandra/Scylla and how to retrieve them. Finally, I will talk about how MySQL replication compares with Cassandra replication. Hopefully, when I am done you will be able to relate to Cassandra operations in a practical and useful way.
Scylla Summit 2017 Keynote: NextGen NoSQL with CEO Dor LaorScyllaDB
ScyllaDB CEO and co-founder Dor Laor shares his vision for Scylla and announces Scylla 2.0, a big step towards the first autonomous NoSQL database—one that dynamically tunes itself to varying conditions while always maintaining a high level of performance.
Duarte Nunes presented on distributed materialized views in ScyllaDB. He discussed the challenges of implementing materialized views in a distributed system without a single master, including propagating updates from base tables to views, handling consistency when tables can diverge, and managing concurrent updates safely. His proposed solution uses asynchronous replica-based propagation paired with repair mechanisms and locking or optimistic concurrency to address these issues. Materialized views provide powerful indexing capabilities but also introduce performance overhead that is difficult to avoid given Scylla's data model.
Scylla Summit 2017: A Toolbox for Understanding Scylla in the FieldScyllaDB
In this talk, we will share useful tools and techniques that we are using in the field to understand Scylla clusters. Users will learn how to use those same tools to better understand their deployment.
Some of the questions that will be answered are:
- how to find out which queries are the slowest and why
- how we go about understanding the impact of the data model in a node's performance
- how to check which resources are the bottlenecks in the cluster
Scylla Summit 2017: Performance Evaluation of Scylla as a Database Backend fo...ScyllaDB
JanusGraph, a highly scalable graph database solution, supports historically Cassandra and HBase as database backends. We decided to put Scylla in the mix, certainly searching for the best performing backend. We ran test scenarios that cover high volume reads and writes. In this talk, we will show you the performance results of Scylla vs others and also share our lessons learned during the performance evaluation.
Scylla Summit 2017: Saving Thousands by Running Scylla on EC2 Spot InstancesScyllaDB
Scylla and Spotinst together provide a strong combination of extreme performance and cost reduction. In this talk, we will present how a Scylla cluster can be used on AWS’s EC2 Spot without losing consistency with the help of Spotinst prediction technology and advanced stateful features. We will show a live demo on how to run Scylla on the Spotinst platform.
Scylla Summit 2017: Repair, Backup, Restore: Last Thing Before You Go to Prod...ScyllaDB
Benchmarks are fun to do but when going to production, all sorts of things can happen: anything from hardware outages to human error bringing your database down. Even in a healthy database, a lot of maintenance operations have to periodically run. Do you have the tools necessary to make sure you are good to go?
Scylla Summit 2017: Welcome and Keynote - Nextgen NoSQLScyllaDB
Our CEO and co-founder Dor Laor and our chairman Benny Schnaider sharing their vision for Scylla. This was also our opportunity to announce Scylla 2.0. Our latest release is a big step toward the first autonomous NoSQL database—one that dynamically tunes itself to varying conditions while always maintaining a high level of performance.
Scylla Summit 2017: How to Optimize and Reduce Inter-DC Network Traffic and S...ScyllaDB
The document appears to be a presentation on optimizing inter-data center communication. It discusses key topics like what inter-data center communication involves, the costs associated with it, best practices for setting snitches, keyspaces, client drivers and consistency levels for queries to optimize performance between data centers. It recommends using network topology replication strategies over simple strategies for multi-region deployments, setting load balancing and consistency levels appropriately in clients, and enabling internode compression to reduce costs of communication between data centers. The presentation encourages reviewing client locations, data access patterns, who is reading/writing data, and having conversations between operations and development teams to determine the best use cases.
mParticle's Journey to Scylla from CassandraScyllaDB
mParticle processes 50 billion monthly messages and needed a data store that provides full availability and performance. They previously used Cassandra but faced issues with high latency, complicated tuning, and backlogs of up to 20 hours. They tested Scylla and found it provided significantly lower latency and compaction backlogs with minimal tuning needed. Scylla also offered knowledgeable support. mParticle migrated their data from Cassandra to Scylla, which immediately kept up with their data loads with little to no backlog.
Scylla Summit 2016: Keynote - Big Data Goes NativeScyllaDB
This document discusses Scylla, a new database that aims to improve upon existing databases. It notes several key differences in Scylla's architecture that allow it to be faster and more scalable than other databases, including its use of techniques like log-structured merge trees, lock-free design, and asynchronous programming. The document also outlines Scylla's value proposition as the fastest database with the best high availability and ease of management compared to other options.
How to Monitor and Size Workloads on AWS i3 instancesScyllaDB
There is a new class of machines in town! Amazon recently unveiled i3, a new class of machines targeted at I/O-intensive workloads. Scylla will officially support i3, and previews are already available.
Join our webinar to learn how to build a state-of-the-art database solution. Presenters Glauber Costa and Eyal Gutkind will cover how to:
- Determine which workloads can benefit from i3 instances
- Ensure Scylla fully leverages the great resources in the i3 family
- Effectively navigate the Scylla monitoring system and identify bottlenecks
You'll also see a live demonstration with a dashboard featuring an i3 cluster with different data models and workloads.
Scylla Summit 2017: Stretching Scylla Silly: The Datastore of a Graph Databas...ScyllaDB
In this talk, we will cover the lay of the land of graph databases. We will talk about what it takes to run a highly available hosted solution in the cloud while giving users a seamless vertical and horizontal scaling solution, and share our experiences migrating from an Apache Cassandra backed graphDB as-a-service solution.
How to achieve no compromise performance and availabilityScyllaDB
ScyllaDB co-founders Dor Laor and Avi Kivity discuss why they started ScyllaDB, the decision decisions they made to achieve no-compromise performance and availability, and give a demo on how to get up and running on Docker.
Scylla Summit 2017: Cry in the Dojo, Laugh in the Battlefield: How We Constan...ScyllaDB
Testing a complex system like Scylla is a challenge on its own. There are many environments, workloads, and problems. Simple problems become increasingly worse at scale. In this talk, we will explore the testing method that we employ in our QA lab and our plans to make it even better in years to come.
Scylla Summit 2017: Gocqlx - A Productivity Toolkit for Scylla and Apache Cas...ScyllaDB
The document describes how to use gocqlx to interact with Cassandra databases. It defines a Tweet struct to map to a Cassandra table and shows examples of using gocqlx to insert and select tweets, including building queries, binding parameters, and executing queries. Benchmark results are shown that demonstrate gocqlx performing inserts and selections faster than raw gocql.
The document discusses optimizing Oracle and Siebel applications on the Sun UltraSPARC T1 platform. It describes how Siebel's multi-threaded architecture is well-suited to the T1 processor's ability to run multiple threads in parallel. It provides examples of consolidating Siebel environments and optimizing performance through Solaris, Siebel, and Oracle database tuning. Metrics show Siebel performing well with low CPU utilization on T1 systems.
How we got to 1 millisecond latency in 99% under repair, compaction, and flus...ScyllaDB
Scylla is an open source reimplementation of Cassandra which performs up to 10X with drop in-replacement compatibility. At ScyllaDB, performance matters but even more importantly, stable performance under any circumstances.
A key factor for our consistent performance is our reliance on userspace schedulers. Scheduling in userspace allows the application, the database in our case to have better control on the different priorities each task has and to provide an SLA to selected operations. Scylla used to have an I/O scheduler and recently won a CPU scheduler.
At ScyllaDB, we make architectural decisions that provide not only low latencies but consistently low latencies at higher percentiles. This begins with our choice of language and key architectural decisions such as not using the Linux page-cache, and is fulfilled by autonomous database control, a set of algorithms, which guarantees that the system will adapt to changes in the workload. In the last year, we have made changes to Scylla that provide latencies that are consistent in every percentile. In this talk, Dor Laor will recap those changes and discuss what ScyllaDB is doing in the future.
The document discusses optimizing Oracle and Siebel applications on Sun Microsystems' UltraSPARC T1 (Niagara) platform. It provides an overview of Siebel architecture and its suitability for the T1 processor. Performance benchmarks show Siebel scaling well by taking advantage of the T1 processor's multithreading capabilities. The document also discusses various optimizations that can be done at the application, database, storage, and operating system levels to further improve performance.
Running Apache Spark on a High-Performance Cluster Using RDMA and NVMe Flash ...Databricks
This document summarizes a presentation about using the Crail distributed storage system to improve Spark performance on high-performance computing clusters with RDMA networking and NVMe flash storage. The key points are:
1) Traditional Spark storage and networking APIs do not bypass the operating system kernel, limiting performance on modern hardware.
2) The Crail system provides user-level APIs for RDMA networking and NVMe flash to improve Spark shuffle, join, and sorting workloads by 2-10x on a 128-node cluster.
3) Crail allows Spark workloads to fully utilize high-speed networks and disaggregate memory and flash storage across nodes without performance penalties.
The document discusses using flash storage to accelerate application performance. It describes how flash provides faster data transfer rates, IOPS and lower latency compared to HDDs. It outlines different ways flash can be used, including as a host-side PCIe device, array-based caching, or within an all-flash array optimized for flash. The Whiptail storage system is highlighted as providing high throughput, IOPS and endurance while reducing power, space and cooling needs compared to HDD solutions. It can support multiple workloads on a single system.
This document discusses disk I/O performance testing tools. It introduces SQLIO and IOMETER for measuring disk throughput, latency, and IOPS. Examples are provided for running SQLIO tests and interpreting the output, including metrics like throughput in MB/s, latency in ms, and I/O histograms. Other disk performance factors discussed include the number of outstanding I/Os, block size, and sequential vs random access patterns.
This document discusses various computer storage technologies including:
- FIFO and LRU caching algorithms.
- Hard disk drives including cylinders, tracks, sectors, and clusters. Latency is discussed in relation to rotational speed.
- Solid state drives and their advantages over hard disk drives like speed and lack of moving parts.
- SATA vs ATA interfaces and performance comparisons.
- RAID disk arrays and their use of redundancy to increase reliability.
- NTFS and FAT16 file systems. NTFS supports long filenames and compression while FAT16 has limitations like a 2GB size limit.
Nachos 2
The document discusses various data storage technologies including FIFO, LRU, cache memory, hard disk drives, solid state drives, SATA vs ATA interfaces, and RAID disk arrays. It provides details on the characteristics and implementations of each technology, such as how FIFO and LRU ordering techniques work, the components and operation of hard disks, performance comparisons of SATA and ATA interfaces, and the use of redundancy in RAID arrays.
Veracity's Coldstore Arcus - Storage as the foundation of your surveillance s...Alex Kwan
The document describes Coldstore, a storage solution from Veracity that addresses challenges with traditional hard disk drives (HDDs) in surveillance systems. It presents SFS (Sequential Filing System) and LAID (Linear Array of Idle Disks) which utilize HDDs more efficiently for surveillance needs by writing data sequentially and only powering one disk at a time. This reduces wear, heat, vibration and failure rates compared to RAID solutions. When an HDD fails in LAID, the system skips it and continues functioning normally without data loss or rebuild times.
The document discusses new hardware and software from Oracle. It highlights several new Oracle server systems including the SPARC T5-8, M6-32, and T5-2. It summarizes their leading benchmark performance results for SPECjEnterprise, TPC-H, TPC-C, and SPECjbb2013. It also discusses new features of Oracle Solaris 11 including predictive self-healing, encryption, and improvements for Oracle RAC databases.
This document discusses various options for deploying solid state drives (SSDs) in the data center to address storage performance issues. It describes all-flash arrays that use only SSDs, hybrid arrays that combine SSDs and hard disk drives, and server-side flash caching. Key points covered include the performance benefits of SSDs over HDDs, different types of SSDs, form factors, deployment architectures like all-flash arrays from vendors, hybrid arrays, server-side caching software, virtual storage appliances, and hyperconverged infrastructure systems. Choosing the best solution depends on factors like performance needs, capacity, data services required, and budget.
P99 Pursuit: 8 Years of Battling P99 LatencyScyllaDB
Performance engineering is a Sisyphean hill climb for perfection. Those who climb the hill are hardly ever satisfied with the results. You should always ask yourself where the bottleneck is today and what’s holding you back. Great performance improves your software. It enables you to run fewer layers, manage 10x less machines, simplifies your stack, and more.
In this keynote session, ScyllaDB CEO Dor Laor will cover the principles for successful creation of projects like ScyllaDB, KVM, the Linux kernel and explain why they spurred his vision for the P99 CONF.
This document discusses techniques for implementing storage tiering to simplify management, lower costs, and increase performance. It describes using IBM's Easy Tier technology to automatically move data between tiers of flash, disk, and tape storage based on I/O density and age. The tiers include flash, solid state drives, enterprise HDDs, and nearline HDDs. Easy Tier measures activity every 5 minutes and moves hot data to faster tiers and cold data to slower tiers with little administration needed. Case studies show how storage tiering saved IBM Global Accounts $17 million in one year and $90 million over 5 years by optimizing data placement across tiers.
Aurora is Amazon's cloud database that provides enterprise-grade capabilities at lower costs than traditional databases. It offers speed and availability through a distributed, fault-tolerant storage system and automatic scaling of storage and compute resources. Aurora provides cross-region replication for high availability and data locality. Engineering Aurora requires experience in databases, storage systems, and distributed systems.
Nimble Storage is developing flash-enabled storage solutions including an accelerator appliance and storage server. The accelerator increases storage cache by 100x and reduces latency by 25x compared to disk-only solutions. It is targeted at the $15 billion networked storage market. Nimble's technology utilizes a new flash-optimized file system and compression to provide compelling price/performance advantages over competitors. The company is led by experienced engineers from Data Domain, NetApp, and other storage firms.
It’s been an exciting year for Amazon Aurora, the MySQL-compatible relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. In this deep dive session, we’ll discuss best practices and explore new features, include high availability options and new integrations with AWS services. We’ll also discuss the recently-announced Aurora with PostgreSQL compatibility.
This document provides an overview and update on Amazon Aurora, Amazon's relational database service. It discusses new performance enhancements including improved read performance through caching, NUMA-aware scheduling, and lock compression to reduce contention. New availability features are also summarized, such as automatic repair and replacement of failed database nodes and storage volumes that can grow to 64TB. The document outlines Aurora's architecture advantages over traditional databases for scaling in the cloud through its distributed, self-healing design.
RDFox is an optimized RDF triple store and parallel Datalog reasoner that supports OWL 2 RL. It features low memory consumption per triple, high scalability with CPU threads, and integration as a Java or C++ library. Evaluation shows RDFox achieves near-linear speedup on multi-core systems and outperforms competitors for query answering large datasets.
Measuring Database Performance on Bare Metal AWS InstancesScyllaDB
AWS has recently announced a new type of instance targeted at I/O intensive applications, the i3.metal. That instance does away with the virtualization layer altogether and gives back the resources that would otherwise be used by the hypervisor back to the application.
To use all of those resources — 72 CPUs and 512GB of memory — a database needs to be have the ability to scale both up and out.
In this webinar we will look into the performance of Scylla running in a few of those instances versus Apache Cassandra running in their sweet spot, a larger fleet of smaller instances. We will discuss how much of the gains come from the database design and how much come from the removal of the virtualization layer.
Key takeaways:
How to properly compare two different database technologies while being fair to both
How to choose the optimal setup for your Scylla deployment
How AWS’s bare metal servers enable Scylla users to draw a significant performance boost.
Amazon Aurora is a fully managed relational database engine that combines the speed and availability of high-end commercial databases with the simplicity and cost-effectiveness of open source databases. It is purpose-built for the cloud using a new architectural model and distributed systems techniques to provide far higher performance, availability and durability than previously possible using conventional monolithic database architectures. Amazon Aurora packs a lot of innovations in the engine and storage layers. In this session, we will do a deep-dive into some of the key innovations behind Amazon Aurora, new improvements to Aurora's performance, availability and cost-effectiveness and discuss best practices and optimal configurations.
Similar to Scylla Summit 2017: Intel Optane SSDs as the New Accelerator in Your Data Center (20)
Using ScyllaDB for Real-Time Write-Heavy WorkloadsScyllaDB
Keeping latencies low for highly concurrent, intensive data ingestion
ScyllaDB’s “sweet spot” is workloads over 50K operations per second that require predictably low (e.g., single-digit millisecond) latency. And its unique architecture makes it particularly valuable for the real-time write-heavy workloads such as those commonly found in IoT, logging systems, real-time analytics, and order processing.
Join ScyllaDB technical director Felipe Cardeneti Mendes and principal field engineer, Lubos Kosco to learn about:
- Common challenges that arise with real-time write-heavy workloads
- The tradeoffs teams face and tips for negotiating them
- ScyllaDB architectural elements that support real-time write-heavy workloads
- How your peers are using ScyllaDB with similar workloads
Unconventional Methods to Identify Bottlenecks in Low-Latency and High-Throug...ScyllaDB
In this presentation, we explore how standard profiling and monitoring methods may fall short in identifying bottlenecks in low-latency data ingestion workflows. Instead, we showcase the power of simple yet clever methods that can uncover hidden performance limitations.
Attendees will discover unconventional techniques, including clever logging, targeted instrumentation, and specialized metrics, to pinpoint bottlenecks accurately. Real-world use cases will be presented to demonstrate the effectiveness of these methods. By the end of the session, attendees will be equipped with alternative approaches to identify bottlenecks and optimize their low-latency data ingestion workflows for high throughput.
Mitigating the Impact of State Management in Cloud Stream Processing SystemsScyllaDB
Stream processing is a crucial component of modern data infrastructure, but constructing an efficient and scalable stream processing system can be challenging. Decoupling compute and storage architecture has emerged as an effective solution to these challenges, but it can introduce high latency issues, especially when dealing with complex continuous queries that necessitate managing extra-large internal states.
In this talk, we focus on addressing the high latency issues associated with S3 storage in stream processing systems that employ a decoupled compute and storage architecture. We delve into the root causes of latency in this context and explore various techniques to minimize the impact of S3 latency on stream processing performance. Our proposed approach is to implement a tiered storage mechanism that leverages a blend of high-performance and low-cost storage tiers to reduce data movement between the compute and storage layers while maintaining efficient processing.
Throughout the talk, we will present experimental results that demonstrate the effectiveness of our approach in mitigating the impact of S3 latency on stream processing. By the end of the talk, attendees will have gained insights into how to optimize their stream processing systems for reduced latency and improved cost-efficiency.
Measuring the Impact of Network Latency at TwitterScyllaDB
Widya Salim and Victor Ma will outline the causal impact analysis, framework, and key learnings used to quantify the impact of reducing Twitter's network latency.
Architecting a High-Performance (Open Source) Distributed Message Queuing Sys...ScyllaDB
BlazingMQ is a new open source* distributed message queuing system developed at and published by Bloomberg. It provides highly-performant queues to applications for asynchronous, efficient, and reliable communication. This system has been used at scale at Bloomberg for eight years, where it moves terabytes of data and billions of messages across tens of thousands of queues in production every day.
BlazingMQ provides highly-available, fault-tolerant queues courtesy of replication based on the Raft consensus algorithm. In addition, it provides a rich set of enterprise message routing strategies, enabling users to implement a variety of scenarios for message processing.
Written in C++ from the ground up, BlazingMQ has been architected with low latency as one of its core requirements. This has resulted in some unique design and implementation choices at all levels of the system, such as its lock-free threading model, custom memory allocators, compact wire protocol, multi-hop network topology, and more.
This talk will provide an overview of BlazingMQ. We will then delve into the system’s core design principles, architecture, and implementation details in order to explore the crucial role they play in its performance and reliability.
*BlazingMQ will be released as open source between now and P99 (exact timing is still TBD)
Noise Canceling RUM by Tim Vereecke, AkamaiScyllaDB
Noisy Real User Monitoring (RUM) data can ruin your P99!
We introduce a fresh concept called ""Human Visible Navigations"" (HVN) to tackle this risk; we focus on the experiences you actually care about when talking about the speed of our sites:
- Human: We exclude noise coming from bots and synthetic measurements.
- Visible: We remove any partial or fully hidden experiences. These tend to be very slow but users don’t see this slowness.
- Navigations: We ignore lightning fast back-forward navigations which usually have few optimisation opportunities.
Adopting Human Visible Navigations provides you with these key benefits:
- Fewer changes staying below the radar
- Fewer data fluctuations
- Fewer blindspots when finding bottlenecks
- Better correlation with business metrics
This is supported by plenty of real world examples coming from the world's largest scale modeling site (6M Monthly visits) in combination with aggregated data from the brand new rumarchive.com (open source)
After attending this session; your P99 and other percentiles will become less noisy and easier to tune!
Always-on Profiling of All Linux Threads, On-CPU and Off-CPU, with eBPF & Con...ScyllaDB
In this session, Tanel introduces a new open source eBPF tool for efficiently sampling both on-CPU events and off-CPU events for every thread (task) in the OS. Linux standard performance tools (like perf) allow you to easily profile on-CPU threads doing work, but if we want to include the off-CPU timing and reasons for the full picture, things get complicated. Combining eBPF task state arrays with periodic sampling for profiling allows us to get both a system-level overview of where threads spend their time, even when blocked and sleeping, and allow us to drill down into individual thread level, to understand why.
Performance Budgets for the Real World by Tammy EvertsScyllaDB
Performance budgets have been around for more than ten years. Over those years, we’ve learned a lot about what works, what doesn’t, and what we need to improve. In this session, Tammy revisits old assumptions about performance budgets and offers some new best practices. Topics include:
• Understanding performance budgets vs. performance goals
• Aligning budgets with user experience
• Pros and cons of Core Web Vitals
• How to stay on top of your budgets to fight regressions
Using Libtracecmd to Analyze Your Latency and Performance TroublesScyllaDB
Trying to figure out why your application is responding late can be difficult, especially if it is because of interference from the operating system. This talk will briefly go over how to write a C program that can analyze what in the Linux system is interfering with your application. It will use trace-cmd to enable kernel trace events as well as tracing lock functions, and it will then go over a quick tutorial on how to use libtracecmd to read the created trace.dat file to uncover what is the cause of interference to you application.
Reducing P99 Latencies with Generational ZGCScyllaDB
With the low-latency garbage collector ZGC, GC pause times are no longer a big problem in Java. With sub-millisecond pause times there are instead other things in the GC and JVM that can cause application threads to experience unexpected latencies. This talk will dig into a specific use where the GC pauses are no longer the cause of unexpected latencies and look at how adding generations to ZGC help lower the p99 application latencies.
5 Hours to 7.7 Seconds: How Database Tricks Sped up Rust Linting Over 2000XScyllaDB
Linters are a type of database! They are a collection of lint rules — queries that look for rule violations to report — plus a way to execute those queries over a source code dataset.
This is a case study about using database ideas to build a linter that looks for breaking changes in Rust library APIs. Maintainability and performance are key: new Rust releases tend to have mutually-incompatible ways of representing API information, and we cannot afford to reimplement and optimize dozens of rules for each Rust version separately. Fortunately, databases don't require rewriting queries when the underlying storage format or query plan changes! This allows us to ship massive optimizations and support multiple Rust versions without making any changes to the queries that describe lint rules.
Ship now, optimize later"" can be a sustainable development practice after all — join us to see how!
How Netflix Builds High Performance Applications at Global ScaleScyllaDB
We all want to build applications that are blazingly fast. We also want to scale them to users all over the world. Can the two happen together? Can users in the slowest of environments also get a fast experience? Learn how we do this at Netflix: how we understand every user's needs and preferences and build high performance applications that work for every user, every time.
Conquering Load Balancing: Experiences from ScyllaDB DriversScyllaDB
Load balancing seems simple on the surface, with algorithms like round-robin, but the real world loves throwing curveballs. Join me in this session as we delve into the intricacies of load balancing within ScyllaDB Drivers. Discover firsthand experiences from our journey in driver development, where we employed the Power of Two Choices algorithm, optimized the implementation of load balancing in Rust Driver, mitigated cloud costs through zone-aware load balancing and combated the issue of overloading a particular core of ScyllaDB. Be prepared to delve into the practical and theoretical aspects of load balancing, gaining valuable insights along the way.
Interaction Latency: Square's User-Centric Mobile Performance MetricScyllaDB
Mobile performance metrics often take inspiration from the backend world and measure resource usage (CPU usage, memory usage, etc) and workload durations (how long a piece of code takes to run).
However, mobile apps are used by humans and the app performance directly impacts their experience, so we should primarily track user-centric mobile performance metrics. Following the lead of tech giants, the mobile industry at large is now adopting the tracking of app launch time and smoothness (jank during motion).
At Square, our customers spend most of their time in the app long after it's launched, and they don't scroll much, so app launch time and smoothness aren't critical metrics. What should we track instead?
This talk will introduce you to Interaction Latency, a user-centric mobile performance metric inspired from the Web Vital metric Interaction to Next Paint"" (web.dev/inp). We'll go over why apps need to track this, how to properly implement its tracking (it's tricky!), how to aggregate this metric and what thresholds you should target.
How to Avoid Learning the Linux-Kernel Memory ModelScyllaDB
The Linux-kernel memory model (LKMM) is a powerful tool for developing highly concurrent Linux-kernel code, but it also has a steep learning curve. Wouldn't it be great to get most of LKMM's benefits without the learning curve?
This talk will describe how to do exactly that by using the standard Linux-kernel APIs (locking, reference counting, RCU) along with a simple rules of thumb, thus gaining most of LKMM's power with less learning. And the full LKMM is always there when you need it!
99.99% of Your Traces are Trash by Paige CruzScyllaDB
Distributed tracing is still finding its footing in many organizations today, one challenge to overcome is the data volume - keeping 100% of your traces is expensive and unnecessary. Enter sampling - head vs tail how do you decide? Let’s look at the design of Sifter and get familiar with why tail-based sampling is the way to enact a cost-effective tracing solution while actually increasing the system’s observability.
Square's Lessons Learned from Implementing a Key-Value Store with RaftScyllaDB
To put it simply, Raft is used to make a use case (e.g., key-value store, indexing system) more fault tolerant to increase availability using replication (despite server and network failures). Raft has been gaining ground due to its simplicity without sacrificing consistency and performance.
Although we'll cover Raft's building blocks, this is not about the Raft algorithm; it is more about the micro-lessons one can learn from building fault-tolerant, strongly consistent distributed systems using Raft. Things like majority agreement rule (quorum), write-ahead log, split votes & randomness to reduce contention, heartbeats, split-brain syndrome, snapshots & logs replay, client requests dedupe & idempotency, consistency guarantees (linearizability), leases & stale reads, batching & streaming, parallelizing persisting & broadcasting, version control, and more!
And believe it or not, you might be using some of these techniques without even realizing it!
This is inspired by Raft paper (raft.github.io), publications & courses on Raft, and an attempt to implement a key-value store using Raft as a side project.
A Deep Dive Into Concurrent React by Matheus AlbuquerqueScyllaDB
Writing fluid user interfaces becomes more and more challenging as the application complexity increases. In this talk, we’ll explore how proper scheduling improves your app’s experience by diving into some of the concurrent React features, understanding their rationales, and how they work under the hood.
How UiPath Discovery Suite supports identification of Agentic Process Automat...DianaGray10
📚 Understand the basics of the newly persona-based LLM-powered Agentic Process Automation and discover how existing UiPath Discovery Suite products like Communication Mining, Process Mining, and Task Mining can be leveraged to identify APA candidates.
Topics Covered:
💡 Idea Behind APA: Explore the innovative concept of Agentic Process Automation and its significance in modern workflows.
🔄 How APA is Different from RPA: Learn the key differences between Agentic Process Automation and Robotic Process Automation.
🚀 Discover the Advantages of APA: Uncover the unique benefits of implementing APA in your organization.
🔍 Identifying APA Candidates with UiPath Discovery Products: See how UiPath's Communication Mining, Process Mining, and Task Mining tools can help pinpoint potential APA candidates.
🔮 Discussion on Expected Future Impacts: Engage in a discussion on the potential future impacts of APA on various industries and business processes.
Enhance your knowledge on the forefront of automation technology and stay ahead with Agentic Process Automation. 🧠💼✨
Speakers:
Arun Kumar Asokan, Delivery Director (US) @ qBotica and UiPath MVP
Naveen Chatlapalli, Solution Architect @ Ashling Partners and UiPath MVP
Self-Healing Test Automation Framework - HealeniumKnoldus Inc.
Revolutionize your test automation with Healenium's self-healing framework. Automate test maintenance, reduce flakes, and increase efficiency. Learn how to build a robust test automation foundation. Discover the power of self-healing tests. Transform your testing experience.
Retrieval Augmented Generation Evaluation with RagasZilliz
Retrieval Augmented Generation (RAG) enhances chatbots by incorporating custom data in the prompt. Using large language models (LLMs) as judge has gained prominence in modern RAG systems. This talk will demo Ragas, an open-source automation tool for RAG evaluations. Christy will talk about and demo evaluating a RAG pipeline using Milvus and RAG metrics like context F1-score and answer correctness.
Cracking AI Black Box - Strategies for Customer-centric Enterprise ExcellenceQuentin Reul
The democratization of Generative AI is ushering in a new era of innovation for enterprises. Discover how you can harness this powerful technology to deliver unparalleled customer value and securing a formidable competitive advantage in today's competitive market. In this session, you will learn how to:
- Identify high-impact customer needs with precision
- Harness the power of large language models to address specific customer needs effectively
- Implement AI responsibly to build trust and foster strong customer relationships
Whether you're at the early stages of your AI journey or looking to optimize existing initiatives, this session will provide you with actionable insights and strategies needed to leverage AI as a powerful catalyst for customer-driven enterprise success.
The History of Embeddings & Multimodal EmbeddingsZilliz
Frank Liu will walk through the history of embeddings and how we got to the cool embedding models used today. He'll end with a demo on how multimodal RAG is used.
This PDF delves into the aspects of information security from a forensic perspective, focusing on privacy leaks. It provides insights into the methods and tools used in forensic investigations to uncover and mitigate privacy breaches in mobile and cloud environments.
The Zaitechno Handheld Raman Spectrometer is a powerful and portable tool for rapid, non-destructive chemical analysis. It utilizes Raman spectroscopy, a technique that analyzes the vibrational fingerprint of molecules to identify their chemical composition. This handheld instrument allows for on-site analysis of materials, making it ideal for a variety of applications, including:
Material identification: Identify unknown materials, minerals, and contaminants.
Quality control: Ensure the quality and consistency of raw materials and finished products.
Pharmaceutical analysis: Verify the identity and purity of pharmaceutical compounds.
Food safety testing: Detect contaminants and adulterants in food products.
Field analysis: Analyze materials in the field, such as during environmental monitoring or forensic investigations.
The Zaitechno Handheld Raman Spectrometer is easy to use and features a user-friendly interface. It is compact and lightweight, making it ideal for field applications. With its rapid analysis capabilities, the Zaitechno Handheld Raman Spectrometer can help you improve efficiency and productivity in your research or quality control workflows.
Scylla Summit 2017: Intel Optane SSDs as the New Accelerator in Your Data Center
1. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Intel® Optane™ SSDs and Scylla
Providing the Speed of an In-memory
Database with Persistency
Tomer Sandler and Frank Ober
2. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Tomer Sandler
Solution Architect @ ScyllaDB
2
Data Center Solution Architect @ Intel®
Frank Ober
3. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Agenda
▪ Introduction
▪ Intel® Optane™ SSD DC P4800X
▪ Scylla as an In-Memory Like Solution
▪ How We Knew Optane™ is Going to “Rock”
▪ Setup and Workloads
▪ Results
▪ TCO: Enterprise SSD vs. Intel® Optane™
▪ Summary
3
4. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Introduction
The Challenge
Providing a solution with the performance of an in-memory like
database without compromises on throughput, latency, and data
persistence.
4
5. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Introduction
The Challenge
Providing a solution with the performance of an in-memory like
database without compromises on throughput, latency, and data
persistence.
How...
Using Scylla and Intel® Optane™ SSD DC P4800X to resolve cold-cache
and data persistence challenges.
5
6. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Intel® Optane™ SSD DC
P4800X
7. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
7
8. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
8
9. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
9
10. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
10
11. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
11
12. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
12
13. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Scylla as an
In-Memory Like
Solution
14. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Scylla as an In-Memory Like Solution
▪ In-Memory Database Requirements
o Sub-millisecond response time
o High throughput
o Support large number of clients concurrently
14
15. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Scylla as an In-Memory Like Solution
▪ In-Memory Database Requirements
o Sub-millisecond response time
o High throughput
o Support large number of clients concurrently
▪ In-Memory Database Challenges
o Cold cache and long warmup times
o Persistency and high availability
o Scalability
o Simplistic data models
15
16. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Scylla as an In-Memory Like Solution
▪ Scylla provides
o Persistent data storage
o High throughput, low latency data access
o Rich data model capabilities
▪ Scylla scales (and scales...)
▪ Scylla needs VERY fast storage media to pair with
▪ Ease fetching and storing information latency
16
17. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
How We Knew
Optane™ is Going
to “Rock”
18. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
How We Knew Optane™ is Going to “Rock”
▪ We used Diskplorer to measure the drives capabilities
o Small wrapper around fio that is used to graph the relationship between
concurrency (I/O depth), throughput, and IOps
18
19. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
How We Knew Optane™ is Going to “Rock”
▪ We used Diskplorer to measure the drives capabilities
o Small wrapper around fio that is used to graph the relationship between
concurrency (I/O depth), throughput, and IOps
o Concurrency is the number of parallel operations that a disk or array can
sustain. With increasing concurrency, the latency increases and we observe
diminishing IOps increases beyond an optimal point
19
20. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
How We Knew Optane™ is Going to “Rock”
▪ We used Diskplorer to measure the drives capabilities
o Small wrapper around fio that is used to graph the relationship between
concurrency (I/O depth), throughput, and IOps
o Concurrency is the number of parallel operations that a disk or array can
sustain. With increasing concurrency, the latency increases and we observe
diminishing IOps increases beyond an optimal point
RandRead test with a 4K buffer:
● Optimal concurrency is ~24
● Throughput: 1.0M IOps
● Latency: 18µs
20
21. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Setup and Workloads
22. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Setup and Workloads
▪ 3 Scylla v2.0 RC servers: 2 x 14 Core CPUs, 128GB DRAM, 2 x Intel®
Optane™ SSD DC P4800X
o CPU: Intel® Xeon® CPU E5-2690 v4 @ 2.60GHz
o Storage: RAID-0 on top of 2 Optane™ drives – total of 750GB per server
o Network: 2 bonded 10Gb Intel® x540 NICs. Bonding type: layer3+4
▪ 3 Client servers: 2 x 14 Core CPUs, 128GB DRAM, using the
cassandra-stress tool with a user profile workload
▪ Set the # of IO queues equal to the # of shards
o /etc/scylla.d/io.conf: SEASTAR_IO="--num-io-queues=54
--max-io-requests=432"
22
23. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Setup and Workloads
▪ Cassandra-stress: User defined mode that allows running
performance tests on custom data models, using yaml files for
configuration
▪ Simple K/V schema used to populate ~50% of the storage capacity
▪ Utilizing all of the server’s RAM (128GB), replication factor set to 3
(RF=3), and the consistency level is set to one (CL=ONE)
▪ Tested 1 / 5 / 10 KByte payloads
o Challenge the default 512B sector size
o Max. IOps for each payload, at very low latency for reads
23
24. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Setup and Workloads
▪ Two scenarios for read tests
o Large working set much larger than the RAM capacity. This scenario lowers the
probability of finding a read partition in Scylla’s cache
o Small working set that will create a higher probability of a partition being
cached in Scylla’s memory
▪ Latency measurements
o Cassandra stress client end-to-end latency results
o Scylla-server side latency results (using `nodetool tablehistograms` command)
24
25. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Results
26. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Latency Test Results
26
Payload Size Test Case (RF=3)
Total Requests
per Sec
Cassandra stress 95%
Latency (ms)
Scylla-server 95%
Latency (ms)
Disk Throughput per
Server (GBps)
Load per
Server
1 KB
key:64b
blob:1kb
Write
300M Partitions
(~50% disk space)
Avg: ~196K
Max: 220K
2.0
Avg: ~1.25
Max: 2.65
~65%
Read
Large Spread
(~75% from Disk)
198K 0.7 0.478
Avg: ~1.65
Max: 2.2
~32%
Read
Small Spread
(All in-Memory)
198K 0.4 0.023 None ~15%
5 KB
key:64b
blob:5kb
Write
75M Partitions
(~54% disk space)
Avg: ~166K
Max: 180K
2.8
Avg: ~2.75
Max: 4.2
~65%
Read
Large Spread
(75% from Disk)
168K 0.9 0.405
Avg: ~1.22
Max: 1.84
~36%
Read
Small Spread
(All in-Memory)
168K 0.5 0.0405 None ~18%
27. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Latency Test Results
27
Payload Size Test Case (RF=3)
Total Requests
per Sec
Cassandra stress 95%
Latency (ms)
Scylla-server 95%
Latency (ms)
Disk Throughput per
Server (GBps)
Load per
Server
10 KB
key:64b
blob:10kb
Write
36M Partitions
(~50% disk space)
120K 2.45
Avg: ~3.7
Max: 4.5
~65%
Read
Large Spread 1
(75% from Disk)
120K 1.0 0.398
Avg: ~0.95
Max 1.72
~30%
Read
Large Spread 2
(75% from Disk)
166K 1.2 0.481
Avg: ~1.35
Max: 2.27
~40%
Read
Small Spread
(All in-Memory)
166K
(120K)
0.6
(0.5)
0.063
(0.051)
None ~22%
28. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Throughput Test Results
28
Payload Size Test Case (RF=1)
Total Requests
per Sec
Cassandra stress 95%
Latency (ms)
Cassandra stress
threads per client
Disk Throughput per
Server (GBps)
Load per
Server
128B
key:64b
blob:128b
Write
600M Partitions
(~8% disk space)
Avg: ~1.95M
Max: 3.05M
7.3 520
Avg: ~0.55
Max: 1.12
~95%
Read 300M
Large Spread
(~50% from Disk)
Avg: ~976K
Max: 1.35M
2.5 120
Avg: ~2.3
Max: 4.29
~94%
Read 600M
Large Spread
(~60% from Disk)
Avg: ~771K
Max: 986K
2.95 120
Avg: ~3.35
Max: 4.53
~94%
Read
Small Spread
(All in-Memory)
Avg: ~2.19M
Max: 2.21M
2.6 300 None ~96%
▪ 128B payload with RF and CL = ONE
▪ 12 cassandra-stress instances (each instance populating a different range).
▪ Read large spread test ran twice, once on the full range (600M partitions) and once on half the
range (300M partitions)
29. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Throughput Test Results
29
Payload Size Test Case (RF=1)
Total Requests
per Sec
Cassandra stress 95%
Latency (ms)
Cassandra stress
threads per client
Disk Throughput per
Server (GBps)
Load per
Server
128B
key:64b
blob:128b
Write
600M Partitions
(~8% disk space)
Avg: ~1.95M
Max: 3.05M
7.3 520
Avg: ~0.55
Max: 1.12
~95%
Read 300M
Large Spread
(~50% from Disk)
Avg: ~976K
Max: 1.35M
2.5 120
Avg: ~2.3
Max: 4.29
~94%
Read 600M
Large Spread
(~60% from Disk)
Avg: ~771K
Max: 986K
2.95 120
Avg: ~3.35
Max: 4.53
~94%
Read
Small Spread
(All in-Memory)
Avg: ~2.19M
Max: 2.21M
2.6 300 None ~96%
▪ 128B payload with RF and CL = ONE
▪ 12 cassandra-stress instances (each instance populating a different range)
▪ Read large spread test ran twice, once on the full range (600M partitions) and once on half the
range (300M partitions)
30. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
TCO
31. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
TCO: Enterprise SSD vs. Intel® Optane™
Intel® Optane™ provide great latency results, and is also more than
50% cheaper compared to DRAM or Enterprise SSD configurations
31
32. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
TCO: Enterprise SSD vs. Intel® Optane™
Intel® Optane™ provide great latency results, and is also more than
50% cheaper compared to DRAM or Enterprise SSD configurations
32
33. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
Summary
34. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
What did we learn
▪ Scylla’s C++ per core scaling architecture and unique I/O scheduling can
fully utilize your infrastructure’s potential for running high-throughput
and low latency workloads
▪ Intel® Optane™ and Scylla achieve the performance of an all in-memory
database
▪ Intel® Optane™ and Scylla resolve the cold-cache and data persistence
challenge without compromising on throughput, latency and performance
▪ Data resides on nonvolatile storage
▪ Scylla server’s 95% write/read latency < 0.5msec at 165K requests per sec
▪ TCO: 50% cheaper than an all in-memory solution
34
35. PRESENTATION TITLE ON ONE LINE
AND ON TWO LINES
First and last name
Position, company
THANK YOU
Tomer@scylladb.com
Please stay in touch
Any questions?
Frank.Ober@intel.com
Check our blogs
- Intel Optane Review
- Intel Optane and Scylla