Skip to main content
AI Assist is now on Stack Overflow. Start a chat to get instant answers from across the network. Sign up to save and share your chats.
0 votes
1 answer
59 views

Im trying to make a PySpark connection to Cassandra DB indexed with Geomesa. Searching about it, I noticed that it uses the Geotools spark runtime since there is no optimized runtime for Cassandra. I'...
Totopo's user avatar
  • 1
0 votes
1 answer
151 views

I got the problem while running spark-submit --master spark://localhost:7077 \ --packages com.datastax.spark:spark-cassandra-connector_2.12:3.5.1, \ org.apache.spark:spark-sql-kafka-0-10_2.12:3.5.1 \ -...
melon's user avatar
  • 1
0 votes
2 answers
103 views

[cassandra running from docker windows] and I am running spark from wsl2 spark-shell --packages com.datastax.spark:spark-cassandra-connector_2.12:3.5.1 [its spark-shell after the command above] and ...
Arka Dash's user avatar
0 votes
1 answer
55 views

I am composing these services in separate docker containers all on the same confluent network: broker: image: confluentinc/cp-server:7.4.0 hostname: broker container_name: broker ...
user28291353's user avatar
1 vote
1 answer
154 views

Please help me fix the above errors based on the code I used The proccesing_data.py code is used to process data using spark-streaming import logging from pyspark.sql import SparkSession from ...
Lê Anh Tuấn 291N40's user avatar
0 votes
1 answer
64 views

I'm try to make connection from Pyspark to Cassandra in virtual environment and the services is installed by docker. I've been using the --packages method to solve the dependencies but it seems doesn'...
ramadhani nugraha's user avatar
2 votes
0 answers
37 views

I'm building an application that allows me to use pyspark to combine an oracle and a cassandra table. The cassandra table's count is in a scale of 100s of millions with the oracle one at a few 1000. ...
GrImPeAper's user avatar
0 votes
0 answers
98 views

I'm trying to create a pipeline in streaming that is calling an API using Airflow, then processing it with Kafka and inserting the data into Cassandra using Spark. I'm struggling when inserting data ...
francollado99's user avatar
1 vote
1 answer
64 views

i am trying to register custom code(for map) like below val session: CqlSession = CassandraConnector.apply(spark.sparkContext).openSession() val codecRegistry: MutableCodecRegistry = session....
Shivam Sajwan's user avatar
2 votes
0 answers
96 views

I am using spark-cassandra-connector_2.11 and version 2.5.2 in my scala application and want to register a custom map codec, but facing issues. Is there any way to register it. I did the same thing on ...
Shivam Sajwan's user avatar
0 votes
1 answer
92 views

I'm working airflow into the Docker container on WindowsPC. I have some problems with apache airflow spark submit operator. I want to write data to a remote Cassandra server. When I was using df.write....
zehra özdemir's user avatar
0 votes
1 answer
317 views

I had this error: py4j.protocol.Py4JJavaError: An error occurred while calling o59.start. : java.lang.NoClassDefFoundError: com/datastax/spark/connector/util/Logging at java.base/java.lang....
francollado99's user avatar
0 votes
1 answer
74 views

I understand that when inserting data, tombstones might be created because of existing null values in the columns of the dataframe. To mitigate this issue and minimize tombstones, insertion queries ...
danmo41's user avatar
0 votes
3 answers
953 views

Context : Sorry to bother you with I am struggling to install a Cassandra-Spark connector. My goal is to install it to use Spark-SQL since Cassandra has strong limitations to do requests. I have : ...
laurentg13's user avatar
0 votes
1 answer
315 views

Running the Python code does not connect to Spark and does not create a database in Cassandra either. I have confirmed the services are up on docker and accessible from the PC. I placed the .jar files ...
Gregory Morris's user avatar

15 30 50 per page
1
2 3 4 5
67