azure-cosmosdb-spark
azure-cosmosdb-spark copied to clipboard
Apache Spark Connector for Azure Cosmos DB
I'm using Azure Databricks Runtime 5.4 (includes Apache Spark 2.4.3, Scala 2.11) with azure_cosmosdb_spark_2_4_0_2_11_1_4_0_uber.jar I try to execute many jobs to the same spark cluster Each job connects to the...
Today, the Spark connector supports two modes for bulk ingestion - BulkUpdate and BulkImport. BulkImport replaces the existing document in its entirety if it exists, or creates a new document...
I'm using Azure Data Bricks and trying to read from the change feed with Spark Structured Stream: spark.readStream.format("com.microsoft.azure.cosmosdb.spark").options(**streamingConfig).load() But I'm getting the following error: Data source com.microsoft.azure.cosmosdb.spark does not support...
Can we please get more graphframes examples? 1. Reading from cosmosdb to spark 2. Writing from spark to cosmosdb 3. Simple transformations of vertices and edges in spark and writing...
Hi, Are you planning on adding Scala 2.12 builds in parallel with 2.11 builds as Spark officially supports 2.12 now. thanks
To minimize receiving 429 errors when running Spark queries against Cosmos DB. ## Size Assumptions For this initial iteration, we will make the following assumptions. * X number of Spark...
Make better use of DocumentDB's native capabilities (e.g. aggregations, `ORDER BY`, `LIMIT`, etc.) so a more optimized dataset is returned to Apache Spark. For example: - For cumulative aggregations, DocumentDB...
minimal getting started samples for - Batch write using read from blob - Batch read and export to blob - Stream read and create materialized view - Stream write from...
When RU runs out while saving data to CosmosDB using Spark Cosmsos DB Connector (https://github.com/Azure/azure-cosmosdb-spark), some data saving may be skipped.