Anant Corporation Blog: Our research, knowledge, thoughts, and recommendations about building and managing online business platforms.
In this blog, we discuss the various methods of counting rows offered in Datastax Enterprise. Traditional databases maintain counts as a matter of course, but Cassandra’s architecture makes that more difficult. This architecture is the same thing that provides Cassandra its advantages, being scale-able, distributed, and having fast reads and writes. Since we still need row counts, we can use DSE’s features to get them in various circumstances.Continue reading
In Apache Cassandra Lunch #29: Cassandra & Kubernetes Update, we cover updates regarding Cassandra and Kubernetes after the recent KubeCon event. The live recording of Cassandra Lunch, which includes a more in-depth discussion, is also embedded below in case you were not able to attend live. If you would like to attend Apache Cassandra Lunch live, it is hosted every Wednesday at 12 PM EST. Register here now!Continue reading
In this blog, we discuss how to run an Apache Spark job on our DataStax Astra database. Also, a webinar recording is embedded below if you want to watch a live demo where we use Gitpod, sbt, Scala, and Spark-Submit to run 2 Spark jobs against our DataStax Astra instance. If you missed part 1 of this series: Connect Apache Spark and DataStax Astra, it will be linked below.Continue reading
ETL stands for Extract, Transform, and Load. This describes a process through which data becomes more refined. ETL is one of the main skills that data engineers need to master in order to do their jobs well. It was also the topic of our second ever Data Engineer’s lunch discussion. If you missed it, or just want an overview of available ETL frameworks, keep reading. The live recording of the Data Engineer’s Lunch, which includes a more in-depth discussion, is also embedded below in case you could not attend live. If you would like to attend a Data Engineer’s Lunch live, it is hosted every Monday at 12 PM EST. Register here now!Continue reading
Notebooks are a potentially useful tool for those working with Apache Spark. The Databricks platform, built on Spark, also includes notebooks as a standard feature. As part of an ongoing series of webinars on replicating functionality that is part of the Databricks platform, we will be attempting to connect notebooks to Apache Spark ourselves.Continue reading