Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

bug: PySpark write streams failing #10828

Open
1 task done
christophediprima opened this issue Feb 12, 2025 · 0 comments
Open
1 task done

bug: PySpark write streams failing #10828

christophediprima opened this issue Feb 12, 2025 · 0 comments
Labels
bug Incorrect behavior inside of ibis

Comments

@christophediprima
Copy link

What happened?

I am trying to use the spark backend to read from and write to kafka but with no success.

The to_kafka method fails when calling sql with the expression.

df = self._session.sql(self.compile(expr, params=params, limit=limit))

pyspark.errors.exceptions.connect.AnalysisException: Queries with streaming sources must be executed with writeStream.start();

I am not sure you can createOrReplaceTempView out of a stream dataframe and query it with spark.sql.

What version of ibis are you using?

10.0.0

What backend(s) are you using, if any?

PySpark with connect

Relevant log output

python3.11/site-packages/pyspark/sql/connect/client/core.py", line 1539, in _handle_rpc_error
    raise convert_exception(info, status.message) from None
pyspark.errors.exceptions.connect.AnalysisException: Queries with streaming sources must be executed with writeStream.start();
kafka

Code of Conduct

  • I agree to follow this project's Code of Conduct
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Incorrect behavior inside of ibis
Projects
Status: backlog
Development

No branches or pull requests

1 participant