spark-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Apache Spark (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (SPARK-18127) Add hooks and extension points to Spark
Date Fri, 21 Apr 2017 21:47:04 GMT

    [ https://issues.apache.org/jira/browse/SPARK-18127?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15979447#comment-15979447
] 

Apache Spark commented on SPARK-18127:
--------------------------------------

User 'sameeragarwal' has created a pull request for this issue:
https://github.com/apache/spark/pull/17724

> Add hooks and extension points to Spark
> ---------------------------------------
>
>                 Key: SPARK-18127
>                 URL: https://issues.apache.org/jira/browse/SPARK-18127
>             Project: Spark
>          Issue Type: New Feature
>          Components: SQL
>            Reporter: Srinath
>            Assignee: Sameer Agarwal
>
> As a Spark user I want to be able to customize my spark session. I currently want to
be able to do the following things:
> # I want to be able to add custom analyzer rules. This allows me to implement my own
logical constructs; an example of this could be a recursive operator.
> # I want to be able to add my own analysis checks. This allows me to catch problems with
spark plans early on. An example of this can be some datasource specific checks.
> # I want to be able to add my own optimizations. This allows me to optimize plans in
different ways, for instance when you use a very different cluster (for example a one-node
X1 instance). This supersedes the current {{spark.experimental}} methods
> # I want to be able to add my own planning strategies. This supersedes the current {{spark.experimental}}
methods. This allows me to plan my own physical plan, an example of this would to plan my
own heavily integrated data source (CarbonData for example).
> # I want to be able to use my own customized SQL constructs. An example of this would
supporting my own dialect, or be able to add constructs to the current SQL language. I should
not have to implement a complete parse, and should be able to delegate to an underlying parser.
> # I want to be able to track modifications and calls to the external catalog. I want
this API to be stable. This allows me to do synchronize with other systems.
> This API should modify the SparkSession when the session gets started, and it should
NOT change the session in flight.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org


Mime
View raw message