Eyal Gutkind, Head of Solution Architects, ScyllaDB
Spark and Scylla deployments are a common theme. Executing analytics workloads on transactional data provide insights to the business team. ETL workloads using Spark and Scylla are common too. We cover different workloads we have seen in practice and how we helped optimize both Spark and Scylla deployments to support a smooth and efficient workflow. Best practices we discuss include correctly sizing the Spark and Scylla nodes, tuning partitions sizes, setting connectors concurrency and Spark retry policies. In addition, we will cover ways to use Spark and Scylla in migrations from different data models.
Holden Karau, Developer Advocate, Google
Getting started takes only a few minutes. Scylla has an installer for every major platform. If you get stuck, we’re here to help.