Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Integration with Spark #4296

Open
vajaw opened this issue Apr 4, 2023 · 2 comments
Open

Integration with Spark #4296

vajaw opened this issue Apr 4, 2023 · 2 comments

Comments

@vajaw
Copy link

vajaw commented Apr 4, 2023

Will Carbonata be integrated with Spark in the future? Can Spark version 3.1.2 be integrated with Carbonata

@chenliang613
Copy link
Contributor

Yes, the community is considering spark 3.3

@vajaw
Copy link
Author

vajaw commented Apr 7, 2023

Comparing Spark3.1.1 and Spark3.1.2, the parameter list for the writeAndRead method in the DataSource class has increased from 4 to 5. The link between the two versions of code is as follows:
https://github.com/apache/spark/blob/v3.1.2/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
https://github.com/apache/spark/blob/v3.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSource.scala
Due to modifications to the writeAndRead method of the DataSource class in Spark3.1.2, this will result in issues with integrating Carbondata2.3.0 with Spark3.1.2. The reason for the problem is that CarbonReflectionUtils in carbondata2.3.0 references Spark's writeAndRead method. The link for CarbonReflectionUtils is as follows:
https://github.com/apache/carbondata/blob/branch-2.3/integration/spark/src/main/scala/org/apache/spark/util/CarbonReflectionUtils.scala
Will the future community integrate higher versions of Spark3.1.2 or Spark3.1.x

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants