From issues-return-209041-archive-asf-public=cust-asf.ponee.io@spark.apache.org Thu Nov 29 12:18:45 2018 Return-Path: X-Original-To: archive-asf-public@cust-asf.ponee.io Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by mx-eu-01.ponee.io (Postfix) with SMTP id 6072A18066C for ; Thu, 29 Nov 2018 12:18:45 +0100 (CET) Received: (qmail 25857 invoked by uid 500); 29 Nov 2018 11:18:44 -0000 Mailing-List: contact issues-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list issues@spark.apache.org Received: (qmail 25847 invoked by uid 99); 29 Nov 2018 11:18:44 -0000 Received: from pnap-us-west-generic-nat.apache.org (HELO spamd1-us-west.apache.org) (209.188.14.142) by apache.org (qpsmtpd/0.29) with ESMTP; Thu, 29 Nov 2018 11:18:44 +0000 Received: from localhost (localhost [127.0.0.1]) by spamd1-us-west.apache.org (ASF Mail Server at spamd1-us-west.apache.org) with ESMTP id 2185DC73FC for ; Thu, 29 Nov 2018 11:18:44 +0000 (UTC) X-Virus-Scanned: Debian amavisd-new at spamd1-us-west.apache.org X-Spam-Flag: NO X-Spam-Score: -109.501 X-Spam-Level: X-Spam-Status: No, score=-109.501 tagged_above=-999 required=6.31 tests=[ENV_AND_HDR_SPF_MATCH=-0.5, KAM_ASCII_DIVIDERS=0.8, RCVD_IN_DNSWL_MED=-2.3, SPF_PASS=-0.001, USER_IN_DEF_SPF_WL=-7.5, USER_IN_WHITELIST=-100] autolearn=disabled Received: from mx1-lw-us.apache.org ([10.40.0.8]) by localhost (spamd1-us-west.apache.org [10.40.0.7]) (amavisd-new, port 10024) with ESMTP id oYQ1WN-Mg_MA for ; Thu, 29 Nov 2018 11:18:43 +0000 (UTC) Received: from mailrelay1-us-west.apache.org (mailrelay1-us-west.apache.org [209.188.14.139]) by mx1-lw-us.apache.org (ASF Mail Server at mx1-lw-us.apache.org) with ESMTP id D05F360CE0 for ; Thu, 29 Nov 2018 11:12:00 +0000 (UTC) Received: from jira-lw-us.apache.org (unknown [207.244.88.139]) by mailrelay1-us-west.apache.org (ASF Mail Server at mailrelay1-us-west.apache.org) with ESMTP id 6C5FAE00D4 for ; Thu, 29 Nov 2018 11:12:00 +0000 (UTC) Received: from jira-lw-us.apache.org (localhost [127.0.0.1]) by jira-lw-us.apache.org (ASF Mail Server at jira-lw-us.apache.org) with ESMTP id 3210424DD0 for ; Thu, 29 Nov 2018 11:12:00 +0000 (UTC) Date: Thu, 29 Nov 2018 11:12:00 +0000 (UTC) From: "Liang-Chi Hsieh (JIRA)" To: issues@spark.apache.org Message-ID: In-Reply-To: References: Subject: [jira] [Commented] (SPARK-26215) define reserved keywords after SQL standard MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-JIRA-FingerPrint: 30527f35849b9dde25b450d4833f0394 [ https://issues.apache.org/jira/browse/SPARK-26215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16703035#comment-16703035 ] Liang-Chi Hsieh commented on SPARK-26215: ----------------------------------------- Thanks for pinging me. Is "In Spark SQL, we are too tolerant about non-reserved keywords" meaning that we have too many non-reserved keywords which should be defined as reserved keywords? > define reserved keywords after SQL standard > ------------------------------------------- > > Key: SPARK-26215 > URL: https://issues.apache.org/jira/browse/SPARK-26215 > Project: Spark > Issue Type: Improvement > Components: SQL > Affects Versions: 2.4.0 > Reporter: Wenchen Fan > Priority: Major > > There are 2 kinds of SQL keywords: reserved and non-reserved. Reserved keywords can't be used as identifiers. > In Spark SQL, we are too tolerant about non-reserved keywors. A lot of keywords are non-reserved and sometimes it cause ambiguity (IIRC we hit a problem when improving the INTERVAL syntax). > I think it will be better to just follow other databases or SQL standard to define reserved keywords, so that we don't need to think very hard about how to avoid ambiguity. > For reference: https://www.postgresql.org/docs/8.1/sql-keywords-appendix.html -- This message was sent by Atlassian JIRA (v7.6.3#76005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org For additional commands, e-mail: issues-help@spark.apache.org