1

This is a challenge specific to SPARK-SQL and I'm unable to apply two highlighted answers

I'm writing complex data processing logic in SPARK-SQL.

Here is the process I follow ,

  1. Define case class for a table with all attributes.
  2. Register that as table.
  3. Use SQLContext to query the same.

I'm encountering an issue as Scala allows only 22 parameters whereas my table has 50 columns. Only approach I could think of is to break dataset in such a way that it has 22 parameters and combine them later at the end. It does not look like a clean approach. Is there any better approach to this issue ?

user3279189
  • 1,643
  • 8
  • 22
  • 35
  • 1
    Hi , This problem is slightly different as I'm applying this in SPARK-SQL. Spark does not build against SCALA 2.11 at the moment and creating a class (instead of case class) is not an option as I need to apply aggregate function and SPARK-SQL does not recognize this.. – user3279189 Aug 05 '14 at 10:25
  • http://mail-archives.apache.org/mod_mbox/spark-user/201407.mbox/%3CCAJ9c7gCcRUVvx0UX4ctq20m7=bPzxieP+JNjcsYsuyjcVpC8kQ@mail.gmail.com%3E – user3279189 Aug 05 '14 at 15:53

1 Answers1

2

Switch to Scala 2.11 and the case class field limit is gone. Release notes. Issue.

Chris Martin
  • 30,334
  • 10
  • 78
  • 137
Gangstead
  • 4,152
  • 22
  • 35