Return-Path: X-Original-To: archive-asf-public-internal@cust-asf2.ponee.io Delivered-To: archive-asf-public-internal@cust-asf2.ponee.io Received: from cust-asf.ponee.io (cust-asf.ponee.io [163.172.22.183]) by cust-asf2.ponee.io (Postfix) with ESMTP id 4B04E200CA3 for ; Wed, 3 May 2017 00:00:45 +0200 (CEST) Received: by cust-asf.ponee.io (Postfix) id 49A11160B9D; Tue, 2 May 2017 22:00:45 +0000 (UTC) Delivered-To: archive-asf-public@cust-asf.ponee.io Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by cust-asf.ponee.io (Postfix) with SMTP id E4693160BE8 for ; Wed, 3 May 2017 00:00:40 +0200 (CEST) Received: (qmail 20872 invoked by uid 500); 2 May 2017 22:00:39 -0000 Mailing-List: contact commits-help@spark.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Delivered-To: mailing list commits@spark.apache.org Received: (qmail 18707 invoked by uid 99); 2 May 2017 22:00:38 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 May 2017 22:00:38 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 49727DF9A3; Tue, 2 May 2017 22:00:38 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: marmbrus@apache.org To: commits@spark.apache.org Date: Tue, 02 May 2017 22:01:20 -0000 Message-Id: <0556456112ca4fe0a8da6eb8cf4eef9f@git.apache.org> In-Reply-To: <2812d736e01d48afbc308cebd5a36c0b@git.apache.org> References: <2812d736e01d48afbc308cebd5a36c0b@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [45/51] [partial] spark-website git commit: Add Spark 2.1.1 docs archived-at: Tue, 02 May 2017 22:00:45 -0000 http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/ceil.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/ceil.html b/site/docs/2.1.1/api/R/ceil.html new file mode 100644 index 0000000..745ae17 --- /dev/null +++ b/site/docs/2.1.1/api/R/ceil.html @@ -0,0 +1,126 @@ + +R: Computes the ceiling of the given value + + + + + + + + + +
ceil {SparkR}R Documentation
+ +

Computes the ceiling of the given value

+ +

Description

+ +

Computes the ceiling of the given value. +

+ + +

Usage

+ +
+## S4 method for signature 'Column'
+ceil(x)
+
+## S4 method for signature 'Column'
+ceiling(x)
+
+ceil(x)
+
+ + +

Arguments

+ + + + +
x +

Column to compute on.

+
+ + +

Note

+ +

ceil since 1.5.0 +

+

ceiling since 1.5.0 +

+ + +

See Also

+ +

Other math_funcs: acos, +acos,Column-method; asin, +asin,Column-method; atan2, +atan2,Column-method; atan, +atan,Column-method; bin, +bin, bin,Column-method; +bround, bround, +bround,Column-method; cbrt, +cbrt, cbrt,Column-method; +conv, conv, +conv,Column,numeric,numeric-method; +corr, corr, +corr, corr,Column-method, +corr,SparkDataFrame-method; +cosh, cosh,Column-method; +cos, cos,Column-method; +covar_pop, covar_pop, +covar_pop,characterOrColumn,characterOrColumn-method; +cov, cov, cov, +cov,SparkDataFrame-method, +cov,characterOrColumn-method, +covar_samp, covar_samp, +covar_samp,characterOrColumn,characterOrColumn-method; +expm1, expm1,Column-method; +exp, exp,Column-method; +factorial, +factorial,Column-method; +floor, floor,Column-method; +hex, hex, +hex,Column-method; hypot, +hypot, hypot,Column-method; +log10, log10,Column-method; +log1p, log1p,Column-method; +log2, log2,Column-method; +log, log,Column-method; +pmod, pmod, +pmod,Column-method; rint, +rint, rint,Column-method; +round, round,Column-method; +shiftLeft, shiftLeft, +shiftLeft,Column,numeric-method; +shiftRightUnsigned, +shiftRightUnsigned, +shiftRightUnsigned,Column,numeric-method; +shiftRight, shiftRight, +shiftRight,Column,numeric-method; +sign, sign,Column-method, +signum, signum, +signum,Column-method; sinh, +sinh,Column-method; sin, +sin,Column-method; sqrt, +sqrt,Column-method; tanh, +tanh,Column-method; tan, +tan,Column-method; toDegrees, +toDegrees, +toDegrees,Column-method; +toRadians, toRadians, +toRadians,Column-method; +unhex, unhex, +unhex,Column-method +

+ + +

Examples

+ +
## Not run: ceil(df$c)
+## Not run: ceiling(df$c)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/clearCache.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/clearCache.html b/site/docs/2.1.1/api/R/clearCache.html new file mode 100644 index 0000000..be03890 --- /dev/null +++ b/site/docs/2.1.1/api/R/clearCache.html @@ -0,0 +1,45 @@ + +R: Clear Cache + + + + + + + + + +
clearCache {SparkR}R Documentation
+ +

Clear Cache

+ +

Description

+ +

Removes all cached tables from the in-memory cache. +

+ + +

Usage

+ +
+## Default S3 method:
+clearCache()
+
+ + +

Note

+ +

clearCache since 1.4.0 +

+ + +

Examples

+ +
## Not run: 
+##D clearCache()
+## End(Not run)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/clearJobGroup.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/clearJobGroup.html b/site/docs/2.1.1/api/R/clearJobGroup.html new file mode 100644 index 0000000..c8ce1f5 --- /dev/null +++ b/site/docs/2.1.1/api/R/clearJobGroup.html @@ -0,0 +1,46 @@ + +R: Clear current job group ID and its description + + + + + + + + + +
clearJobGroup {SparkR}R Documentation
+ +

Clear current job group ID and its description

+ +

Description

+ +

Clear current job group ID and its description +

+ + +

Usage

+ +
+## Default S3 method:
+clearJobGroup()
+
+ + +

Note

+ +

clearJobGroup since 1.5.0 +

+ + +

Examples

+ +
## Not run: 
+##D sparkR.session()
+##D clearJobGroup()
+## End(Not run)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/coalesce.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/coalesce.html b/site/docs/2.1.1/api/R/coalesce.html new file mode 100644 index 0000000..7354d10 --- /dev/null +++ b/site/docs/2.1.1/api/R/coalesce.html @@ -0,0 +1,336 @@ + +R: Coalesce + + + + + + + + + +
coalesce {SparkR}R Documentation
+ +

Coalesce

+ +

Description

+ +

Returns a new SparkDataFrame that has exactly numPartitions partitions. +This operation results in a narrow dependency, e.g. if you go from 1000 partitions to 100 +partitions, there will not be a shuffle, instead each of the 100 new partitions will claim 10 of +the current partitions. If a larger number of partitions is requested, it will stay at the +current number of partitions. +

+

Returns the first column that is not NA, or NA if all inputs are. +

+ + +

Usage

+ +
+## S4 method for signature 'SparkDataFrame'
+coalesce(x, numPartitions)
+
+## S4 method for signature 'Column'
+coalesce(x, ...)
+
+coalesce(x, ...)
+
+ + +

Arguments

+ + + + + + + + +
x +

a Column or a SparkDataFrame.

+
numPartitions +

the number of partitions to use.

+
... +

additional argument(s). If x is a Column, additional Columns can be optionally +provided.

+
+ + +

Details

+ +

However, if you're doing a drastic coalesce on a SparkDataFrame, e.g. to numPartitions = 1, +this may result in your computation taking place on fewer nodes than +you like (e.g. one node in the case of numPartitions = 1). To avoid this, +call repartition. This will add a shuffle step, but means the +current upstream partitions will be executed in parallel (per whatever +the current partitioning is). +

+ + +

Note

+ +

coalesce(SparkDataFrame) since 2.1.1 +

+

coalesce(Column) since 2.1.1 +

+ + +

See Also

+ +

repartition +

+

Other SparkDataFrame functions: $, +$,SparkDataFrame-method, $<-, +$<-,SparkDataFrame-method, +select, select, +select,SparkDataFrame,Column-method, +select,SparkDataFrame,character-method, +select,SparkDataFrame,list-method; +SparkDataFrame-class; [, +[,SparkDataFrame-method, [[, +[[,SparkDataFrame,numericOrcharacter-method, +[[<-, +[[<-,SparkDataFrame,numericOrcharacter-method, +subset, subset, +subset,SparkDataFrame-method; +agg, agg, agg, +agg,GroupedData-method, +agg,SparkDataFrame-method, +summarize, summarize, +summarize, +summarize,GroupedData-method, +summarize,SparkDataFrame-method; +arrange, arrange, +arrange, +arrange,SparkDataFrame,Column-method, +arrange,SparkDataFrame,character-method, +orderBy,SparkDataFrame,characterOrColumn-method; +as.data.frame, +as.data.frame,SparkDataFrame-method; +attach, +attach,SparkDataFrame-method; +cache, cache, +cache,SparkDataFrame-method; +collect, collect, +collect,SparkDataFrame-method; +colnames, colnames, +colnames,SparkDataFrame-method, +colnames<-, colnames<-, +colnames<-,SparkDataFrame-method, +columns, columns, +columns,SparkDataFrame-method, +names, +names,SparkDataFrame-method, +names<-, +names<-,SparkDataFrame-method; +coltypes, coltypes, +coltypes,SparkDataFrame-method, +coltypes<-, coltypes<-, +coltypes<-,SparkDataFrame,character-method; +count,SparkDataFrame-method, +nrow, nrow, +nrow,SparkDataFrame-method; +createOrReplaceTempView, +createOrReplaceTempView, +createOrReplaceTempView,SparkDataFrame,character-method; +crossJoin, +crossJoin,SparkDataFrame,SparkDataFrame-method; +dapplyCollect, dapplyCollect, +dapplyCollect,SparkDataFrame,function-method; +dapply, dapply, +dapply,SparkDataFrame,function,structType-method; +describe, describe, +describe, +describe,SparkDataFrame,ANY-method, +describe,SparkDataFrame,character-method, +describe,SparkDataFrame-method, +summary, summary, +summary,SparkDataFrame-method; +dim, +dim,SparkDataFrame-method; +distinct, distinct, +distinct,SparkDataFrame-method, +unique, +unique,SparkDataFrame-method; +dropDuplicates, +dropDuplicates, +dropDuplicates,SparkDataFrame-method; +dropna, dropna, +dropna,SparkDataFrame-method, +fillna, fillna, +fillna,SparkDataFrame-method, +na.omit, na.omit, +na.omit,SparkDataFrame-method; +drop, drop, +drop, drop,ANY-method, +drop,SparkDataFrame-method; +dtypes, dtypes, +dtypes,SparkDataFrame-method; +except, except, +except,SparkDataFrame,SparkDataFrame-method; +explain, explain, +explain,SparkDataFrame-method; +filter, filter, +filter,SparkDataFrame,characterOrColumn-method, +where, where, +where,SparkDataFrame,characterOrColumn-method; +first, first, +first, +first,SparkDataFrame-method, +first,characterOrColumn-method; +gapplyCollect, gapplyCollect, +gapplyCollect, +gapplyCollect,GroupedData-method, +gapplyCollect,SparkDataFrame-method; +gapply, gapply, +gapply, +gapply,GroupedData-method, +gapply,SparkDataFrame-method; +getNumPartitions, +getNumPartitions,SparkDataFrame-method; +groupBy, groupBy, +groupBy,SparkDataFrame-method, +group_by, group_by, +group_by,SparkDataFrame-method; +head, +head,SparkDataFrame-method; +histogram, +histogram,SparkDataFrame,characterOrColumn-method; +insertInto, insertInto, +insertInto,SparkDataFrame,character-method; +intersect, intersect, +intersect,SparkDataFrame,SparkDataFrame-method; +isLocal, isLocal, +isLocal,SparkDataFrame-method; +join, +join,SparkDataFrame,SparkDataFrame-method; +limit, limit, +limit,SparkDataFrame,numeric-method; +merge, merge, +merge,SparkDataFrame,SparkDataFrame-method; +mutate, mutate, +mutate,SparkDataFrame-method, +transform, transform, +transform,SparkDataFrame-method; +ncol, +ncol,SparkDataFrame-method; +persist, persist, +persist,SparkDataFrame,character-method; +printSchema, printSchema, +printSchema,SparkDataFrame-method; +randomSplit, randomSplit, +randomSplit,SparkDataFrame,numeric-method; +rbind, rbind, +rbind,SparkDataFrame-method; +registerTempTable, +registerTempTable, +registerTempTable,SparkDataFrame,character-method; +rename, rename, +rename,SparkDataFrame-method, +withColumnRenamed, +withColumnRenamed, +withColumnRenamed,SparkDataFrame,character,character-method; +repartition, repartition, +repartition,SparkDataFrame-method; +sample, sample, +sample,SparkDataFrame,logical,numeric-method, +sample_frac, sample_frac, +sample_frac,SparkDataFrame,logical,numeric-method; +saveAsParquetFile, +saveAsParquetFile, +saveAsParquetFile,SparkDataFrame,character-method, +write.parquet, write.parquet, +write.parquet,SparkDataFrame,character-method; +saveAsTable, saveAsTable, +saveAsTable,SparkDataFrame,character-method; +saveDF, saveDF, +saveDF,SparkDataFrame,character-method, +write.df, write.df, +write.df, +write.df,SparkDataFrame-method; +schema, schema, +schema,SparkDataFrame-method; +selectExpr, selectExpr, +selectExpr,SparkDataFrame,character-method; +showDF, showDF, +showDF,SparkDataFrame-method; +show, show, +show,Column-method, +show,GroupedData-method, +show,SparkDataFrame-method, +show,WindowSpec-method; +storageLevel, +storageLevel,SparkDataFrame-method; +str, +str,SparkDataFrame-method; +take, take, +take,SparkDataFrame,numeric-method; +union, union, +union,SparkDataFrame,SparkDataFrame-method, +unionAll, unionAll, +unionAll,SparkDataFrame,SparkDataFrame-method; +unpersist, unpersist, +unpersist,SparkDataFrame-method; +withColumn, withColumn, +withColumn,SparkDataFrame,character-method; +with, +with,SparkDataFrame-method; +write.jdbc, write.jdbc, +write.jdbc,SparkDataFrame,character,character-method; +write.json, write.json, +write.json,SparkDataFrame,character-method; +write.orc, write.orc, +write.orc,SparkDataFrame,character-method; +write.text, write.text, +write.text,SparkDataFrame,character-method +

+

Other normal_funcs: Column-class, +column, column, +column, +column,character-method, +column,jobj-method; abs, +abs,Column-method; +bitwiseNOT, bitwiseNOT, +bitwiseNOT,Column-method; +expr, expr, +expr,character-method; +greatest, greatest, +greatest,Column-method; +ifelse, ifelse,Column-method; +is.nan, is.nan,Column-method, +isnan, isnan, +isnan,Column-method; least, +least, least,Column-method; +lit, lit, +lit,ANY-method; nanvl, +nanvl, nanvl,Column-method; +negate, negate, +negate,Column-method; randn, +randn, randn, +randn,missing-method, +randn,numeric-method; rand, +rand, rand, +rand,missing-method, +rand,numeric-method; struct, +struct, +struct,characterOrColumn-method; +when, when, +when,Column-method +

+ + +

Examples

+ +
## Not run: 
+##D sparkR.session()
+##D path <- "path/to/file.json"
+##D df <- read.json(path)
+##D newDF <- coalesce(df, 1L)
+## End(Not run)
+## Not run: coalesce(df$c, df$d, df$e)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/collect.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/collect.html b/site/docs/2.1.1/api/R/collect.html new file mode 100644 index 0000000..db24703 --- /dev/null +++ b/site/docs/2.1.1/api/R/collect.html @@ -0,0 +1,281 @@ + +R: Collects all the elements of a SparkDataFrame and coerces... + + + + + + + + + +
collect {SparkR}R Documentation
+ +

Collects all the elements of a SparkDataFrame and coerces them into an R data.frame.

+ +

Description

+ +

Collects all the elements of a SparkDataFrame and coerces them into an R data.frame. +

+ + +

Usage

+ +
+## S4 method for signature 'SparkDataFrame'
+collect(x, stringsAsFactors = FALSE)
+
+collect(x, ...)
+
+ + +

Arguments

+ + + + + + + + +
x +

a SparkDataFrame.

+
stringsAsFactors +

(Optional) a logical indicating whether or not string columns +should be converted to factors. FALSE by default.

+
... +

further arguments to be passed to or from other methods.

+
+ + +

Note

+ +

collect since 1.4.0 +

+ + +

See Also

+ +

Other SparkDataFrame functions: $, +$,SparkDataFrame-method, $<-, +$<-,SparkDataFrame-method, +select, select, +select,SparkDataFrame,Column-method, +select,SparkDataFrame,character-method, +select,SparkDataFrame,list-method; +SparkDataFrame-class; [, +[,SparkDataFrame-method, [[, +[[,SparkDataFrame,numericOrcharacter-method, +[[<-, +[[<-,SparkDataFrame,numericOrcharacter-method, +subset, subset, +subset,SparkDataFrame-method; +agg, agg, agg, +agg,GroupedData-method, +agg,SparkDataFrame-method, +summarize, summarize, +summarize, +summarize,GroupedData-method, +summarize,SparkDataFrame-method; +arrange, arrange, +arrange, +arrange,SparkDataFrame,Column-method, +arrange,SparkDataFrame,character-method, +orderBy,SparkDataFrame,characterOrColumn-method; +as.data.frame, +as.data.frame,SparkDataFrame-method; +attach, +attach,SparkDataFrame-method; +cache, cache, +cache,SparkDataFrame-method; +coalesce, coalesce, +coalesce, +coalesce,Column-method, +coalesce,SparkDataFrame-method; +colnames, colnames, +colnames,SparkDataFrame-method, +colnames<-, colnames<-, +colnames<-,SparkDataFrame-method, +columns, columns, +columns,SparkDataFrame-method, +names, +names,SparkDataFrame-method, +names<-, +names<-,SparkDataFrame-method; +coltypes, coltypes, +coltypes,SparkDataFrame-method, +coltypes<-, coltypes<-, +coltypes<-,SparkDataFrame,character-method; +count,SparkDataFrame-method, +nrow, nrow, +nrow,SparkDataFrame-method; +createOrReplaceTempView, +createOrReplaceTempView, +createOrReplaceTempView,SparkDataFrame,character-method; +crossJoin, +crossJoin,SparkDataFrame,SparkDataFrame-method; +dapplyCollect, dapplyCollect, +dapplyCollect,SparkDataFrame,function-method; +dapply, dapply, +dapply,SparkDataFrame,function,structType-method; +describe, describe, +describe, +describe,SparkDataFrame,ANY-method, +describe,SparkDataFrame,character-method, +describe,SparkDataFrame-method, +summary, summary, +summary,SparkDataFrame-method; +dim, +dim,SparkDataFrame-method; +distinct, distinct, +distinct,SparkDataFrame-method, +unique, +unique,SparkDataFrame-method; +dropDuplicates, +dropDuplicates, +dropDuplicates,SparkDataFrame-method; +dropna, dropna, +dropna,SparkDataFrame-method, +fillna, fillna, +fillna,SparkDataFrame-method, +na.omit, na.omit, +na.omit,SparkDataFrame-method; +drop, drop, +drop, drop,ANY-method, +drop,SparkDataFrame-method; +dtypes, dtypes, +dtypes,SparkDataFrame-method; +except, except, +except,SparkDataFrame,SparkDataFrame-method; +explain, explain, +explain,SparkDataFrame-method; +filter, filter, +filter,SparkDataFrame,characterOrColumn-method, +where, where, +where,SparkDataFrame,characterOrColumn-method; +first, first, +first, +first,SparkDataFrame-method, +first,characterOrColumn-method; +gapplyCollect, gapplyCollect, +gapplyCollect, +gapplyCollect,GroupedData-method, +gapplyCollect,SparkDataFrame-method; +gapply, gapply, +gapply, +gapply,GroupedData-method, +gapply,SparkDataFrame-method; +getNumPartitions, +getNumPartitions,SparkDataFrame-method; +groupBy, groupBy, +groupBy,SparkDataFrame-method, +group_by, group_by, +group_by,SparkDataFrame-method; +head, +head,SparkDataFrame-method; +histogram, +histogram,SparkDataFrame,characterOrColumn-method; +insertInto, insertInto, +insertInto,SparkDataFrame,character-method; +intersect, intersect, +intersect,SparkDataFrame,SparkDataFrame-method; +isLocal, isLocal, +isLocal,SparkDataFrame-method; +join, +join,SparkDataFrame,SparkDataFrame-method; +limit, limit, +limit,SparkDataFrame,numeric-method; +merge, merge, +merge,SparkDataFrame,SparkDataFrame-method; +mutate, mutate, +mutate,SparkDataFrame-method, +transform, transform, +transform,SparkDataFrame-method; +ncol, +ncol,SparkDataFrame-method; +persist, persist, +persist,SparkDataFrame,character-method; +printSchema, printSchema, +printSchema,SparkDataFrame-method; +randomSplit, randomSplit, +randomSplit,SparkDataFrame,numeric-method; +rbind, rbind, +rbind,SparkDataFrame-method; +registerTempTable, +registerTempTable, +registerTempTable,SparkDataFrame,character-method; +rename, rename, +rename,SparkDataFrame-method, +withColumnRenamed, +withColumnRenamed, +withColumnRenamed,SparkDataFrame,character,character-method; +repartition, repartition, +repartition,SparkDataFrame-method; +sample, sample, +sample,SparkDataFrame,logical,numeric-method, +sample_frac, sample_frac, +sample_frac,SparkDataFrame,logical,numeric-method; +saveAsParquetFile, +saveAsParquetFile, +saveAsParquetFile,SparkDataFrame,character-method, +write.parquet, write.parquet, +write.parquet,SparkDataFrame,character-method; +saveAsTable, saveAsTable, +saveAsTable,SparkDataFrame,character-method; +saveDF, saveDF, +saveDF,SparkDataFrame,character-method, +write.df, write.df, +write.df, +write.df,SparkDataFrame-method; +schema, schema, +schema,SparkDataFrame-method; +selectExpr, selectExpr, +selectExpr,SparkDataFrame,character-method; +showDF, showDF, +showDF,SparkDataFrame-method; +show, show, +show,Column-method, +show,GroupedData-method, +show,SparkDataFrame-method, +show,WindowSpec-method; +storageLevel, +storageLevel,SparkDataFrame-method; +str, +str,SparkDataFrame-method; +take, take, +take,SparkDataFrame,numeric-method; +union, union, +union,SparkDataFrame,SparkDataFrame-method, +unionAll, unionAll, +unionAll,SparkDataFrame,SparkDataFrame-method; +unpersist, unpersist, +unpersist,SparkDataFrame-method; +withColumn, withColumn, +withColumn,SparkDataFrame,character-method; +with, +with,SparkDataFrame-method; +write.jdbc, write.jdbc, +write.jdbc,SparkDataFrame,character,character-method; +write.json, write.json, +write.json,SparkDataFrame,character-method; +write.orc, write.orc, +write.orc,SparkDataFrame,character-method; +write.text, write.text, +write.text,SparkDataFrame,character-method +

+ + +

Examples

+ +
## Not run: 
+##D sparkR.session()
+##D path <- "path/to/file.json"
+##D df <- read.json(path)
+##D collected <- collect(df)
+##D firstName <- collected[[1]]$name
+## End(Not run)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/coltypes.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/coltypes.html b/site/docs/2.1.1/api/R/coltypes.html new file mode 100644 index 0000000..3eeb076 --- /dev/null +++ b/site/docs/2.1.1/api/R/coltypes.html @@ -0,0 +1,295 @@ + +R: coltypes + + + + + + + + + +
coltypes {SparkR}R Documentation
+ +

coltypes

+ +

Description

+ +

Get column types of a SparkDataFrame +

+

Set the column types of a SparkDataFrame. +

+ + +

Usage

+ +
+## S4 method for signature 'SparkDataFrame'
+coltypes(x)
+
+## S4 replacement method for signature 'SparkDataFrame,character'
+coltypes(x) <- value
+
+coltypes(x)
+
+coltypes(x) <- value
+
+ + +

Arguments

+ + + + + + +
x +

A SparkDataFrame

+
value +

A character vector with the target column types for the given +SparkDataFrame. Column types can be one of integer, numeric/double, character, logical, or NA +to keep that column as-is.

+
+ + +

Value

+ +

value A character vector with the column types of the given SparkDataFrame +

+ + +

Note

+ +

coltypes since 1.6.0 +

+

coltypes<- since 1.6.0 +

+ + +

See Also

+ +

Other SparkDataFrame functions: $, +$,SparkDataFrame-method, $<-, +$<-,SparkDataFrame-method, +select, select, +select,SparkDataFrame,Column-method, +select,SparkDataFrame,character-method, +select,SparkDataFrame,list-method; +SparkDataFrame-class; [, +[,SparkDataFrame-method, [[, +[[,SparkDataFrame,numericOrcharacter-method, +[[<-, +[[<-,SparkDataFrame,numericOrcharacter-method, +subset, subset, +subset,SparkDataFrame-method; +agg, agg, agg, +agg,GroupedData-method, +agg,SparkDataFrame-method, +summarize, summarize, +summarize, +summarize,GroupedData-method, +summarize,SparkDataFrame-method; +arrange, arrange, +arrange, +arrange,SparkDataFrame,Column-method, +arrange,SparkDataFrame,character-method, +orderBy,SparkDataFrame,characterOrColumn-method; +as.data.frame, +as.data.frame,SparkDataFrame-method; +attach, +attach,SparkDataFrame-method; +cache, cache, +cache,SparkDataFrame-method; +coalesce, coalesce, +coalesce, +coalesce,Column-method, +coalesce,SparkDataFrame-method; +collect, collect, +collect,SparkDataFrame-method; +colnames, colnames, +colnames,SparkDataFrame-method, +colnames<-, colnames<-, +colnames<-,SparkDataFrame-method, +columns, columns, +columns,SparkDataFrame-method, +names, +names,SparkDataFrame-method, +names<-, +names<-,SparkDataFrame-method; +count,SparkDataFrame-method, +nrow, nrow, +nrow,SparkDataFrame-method; +createOrReplaceTempView, +createOrReplaceTempView, +createOrReplaceTempView,SparkDataFrame,character-method; +crossJoin, +crossJoin,SparkDataFrame,SparkDataFrame-method; +dapplyCollect, dapplyCollect, +dapplyCollect,SparkDataFrame,function-method; +dapply, dapply, +dapply,SparkDataFrame,function,structType-method; +describe, describe, +describe, +describe,SparkDataFrame,ANY-method, +describe,SparkDataFrame,character-method, +describe,SparkDataFrame-method, +summary, summary, +summary,SparkDataFrame-method; +dim, +dim,SparkDataFrame-method; +distinct, distinct, +distinct,SparkDataFrame-method, +unique, +unique,SparkDataFrame-method; +dropDuplicates, +dropDuplicates, +dropDuplicates,SparkDataFrame-method; +dropna, dropna, +dropna,SparkDataFrame-method, +fillna, fillna, +fillna,SparkDataFrame-method, +na.omit, na.omit, +na.omit,SparkDataFrame-method; +drop, drop, +drop, drop,ANY-method, +drop,SparkDataFrame-method; +dtypes, dtypes, +dtypes,SparkDataFrame-method; +except, except, +except,SparkDataFrame,SparkDataFrame-method; +explain, explain, +explain,SparkDataFrame-method; +filter, filter, +filter,SparkDataFrame,characterOrColumn-method, +where, where, +where,SparkDataFrame,characterOrColumn-method; +first, first, +first, +first,SparkDataFrame-method, +first,characterOrColumn-method; +gapplyCollect, gapplyCollect, +gapplyCollect, +gapplyCollect,GroupedData-method, +gapplyCollect,SparkDataFrame-method; +gapply, gapply, +gapply, +gapply,GroupedData-method, +gapply,SparkDataFrame-method; +getNumPartitions, +getNumPartitions,SparkDataFrame-method; +groupBy, groupBy, +groupBy,SparkDataFrame-method, +group_by, group_by, +group_by,SparkDataFrame-method; +head, +head,SparkDataFrame-method; +histogram, +histogram,SparkDataFrame,characterOrColumn-method; +insertInto, insertInto, +insertInto,SparkDataFrame,character-method; +intersect, intersect, +intersect,SparkDataFrame,SparkDataFrame-method; +isLocal, isLocal, +isLocal,SparkDataFrame-method; +join, +join,SparkDataFrame,SparkDataFrame-method; +limit, limit, +limit,SparkDataFrame,numeric-method; +merge, merge, +merge,SparkDataFrame,SparkDataFrame-method; +mutate, mutate, +mutate,SparkDataFrame-method, +transform, transform, +transform,SparkDataFrame-method; +ncol, +ncol,SparkDataFrame-method; +persist, persist, +persist,SparkDataFrame,character-method; +printSchema, printSchema, +printSchema,SparkDataFrame-method; +randomSplit, randomSplit, +randomSplit,SparkDataFrame,numeric-method; +rbind, rbind, +rbind,SparkDataFrame-method; +registerTempTable, +registerTempTable, +registerTempTable,SparkDataFrame,character-method; +rename, rename, +rename,SparkDataFrame-method, +withColumnRenamed, +withColumnRenamed, +withColumnRenamed,SparkDataFrame,character,character-method; +repartition, repartition, +repartition,SparkDataFrame-method; +sample, sample, +sample,SparkDataFrame,logical,numeric-method, +sample_frac, sample_frac, +sample_frac,SparkDataFrame,logical,numeric-method; +saveAsParquetFile, +saveAsParquetFile, +saveAsParquetFile,SparkDataFrame,character-method, +write.parquet, write.parquet, +write.parquet,SparkDataFrame,character-method; +saveAsTable, saveAsTable, +saveAsTable,SparkDataFrame,character-method; +saveDF, saveDF, +saveDF,SparkDataFrame,character-method, +write.df, write.df, +write.df, +write.df,SparkDataFrame-method; +schema, schema, +schema,SparkDataFrame-method; +selectExpr, selectExpr, +selectExpr,SparkDataFrame,character-method; +showDF, showDF, +showDF,SparkDataFrame-method; +show, show, +show,Column-method, +show,GroupedData-method, +show,SparkDataFrame-method, +show,WindowSpec-method; +storageLevel, +storageLevel,SparkDataFrame-method; +str, +str,SparkDataFrame-method; +take, take, +take,SparkDataFrame,numeric-method; +union, union, +union,SparkDataFrame,SparkDataFrame-method, +unionAll, unionAll, +unionAll,SparkDataFrame,SparkDataFrame-method; +unpersist, unpersist, +unpersist,SparkDataFrame-method; +withColumn, withColumn, +withColumn,SparkDataFrame,character-method; +with, +with,SparkDataFrame-method; +write.jdbc, write.jdbc, +write.jdbc,SparkDataFrame,character,character-method; +write.json, write.json, +write.json,SparkDataFrame,character-method; +write.orc, write.orc, +write.orc,SparkDataFrame,character-method; +write.text, write.text, +write.text,SparkDataFrame,character-method +

+ + +

Examples

+ +
## Not run: 
+##D irisDF <- createDataFrame(iris)
+##D coltypes(irisDF) # get column types
+## End(Not run)
+## Not run: 
+##D sparkR.session()
+##D path <- "path/to/file.json"
+##D df <- read.json(path)
+##D coltypes(df) <- c("character", "integer") # set column types
+##D coltypes(df) <- c(NA, "numeric") # set column types
+## End(Not run)
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/column.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/column.html b/site/docs/2.1.1/api/R/column.html new file mode 100644 index 0000000..d281921 --- /dev/null +++ b/site/docs/2.1.1/api/R/column.html @@ -0,0 +1,108 @@ + +R: S4 class that represents a SparkDataFrame column + + + + + + + + + +
Column-class {SparkR}R Documentation
+ +

S4 class that represents a SparkDataFrame column

+ +

Description

+ +

The column class supports unary, binary operations on SparkDataFrame columns +

+

Returns a Column based on the given column name. +

+ + +

Usage

+ +
+## S4 method for signature 'jobj'
+column(x)
+
+## S4 method for signature 'character'
+column(x)
+
+column(x)
+
+ + +

Arguments

+ + + + +
x +

Character column name.

+
+ + +

Slots

+ + +
+
jc

reference to JVM SparkDataFrame column

+
+
+ + +

Note

+ +

Column since 1.4.0 +

+

column since 1.6.0 +

+ + +

See Also

+ +

Other normal_funcs: abs, +abs,Column-method; +bitwiseNOT, bitwiseNOT, +bitwiseNOT,Column-method; +coalesce, coalesce, +coalesce, +coalesce,Column-method, +coalesce,SparkDataFrame-method; +expr, expr, +expr,character-method; +greatest, greatest, +greatest,Column-method; +ifelse, ifelse,Column-method; +is.nan, is.nan,Column-method, +isnan, isnan, +isnan,Column-method; least, +least, least,Column-method; +lit, lit, +lit,ANY-method; nanvl, +nanvl, nanvl,Column-method; +negate, negate, +negate,Column-method; randn, +randn, randn, +randn,missing-method, +randn,numeric-method; rand, +rand, rand, +rand,missing-method, +rand,numeric-method; struct, +struct, +struct,characterOrColumn-method; +when, when, +when,Column-method +

+ + +

Examples

+ +
## Not run: column("name")
+
+ + +
[Package SparkR version 2.1.1 Index]
+ http://git-wip-us.apache.org/repos/asf/spark-website/blob/d4f0c34a/site/docs/2.1.1/api/R/columnfunctions.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.1/api/R/columnfunctions.html b/site/docs/2.1.1/api/R/columnfunctions.html new file mode 100644 index 0000000..f286f42 --- /dev/null +++ b/site/docs/2.1.1/api/R/columnfunctions.html @@ -0,0 +1,56 @@ + +R: A set of operations working with SparkDataFrame columns + + + + +
columnfunctions {SparkR}R Documentation
+ +

A set of operations working with SparkDataFrame columns

+ +

Description

+ +

A set of operations working with SparkDataFrame columns +

+ + +

Usage

+ +
+asc(x)
+
+contains(x, ...)
+
+desc(x)
+
+getField(x, ...)
+
+getItem(x, ...)
+
+isNaN(x)
+
+isNull(x)
+
+isNotNull(x)
+
+like(x, ...)
+
+rlike(x, ...)
+
+ + +

Arguments

+ + + + + + +
x +

a Column object.

+
... +

additional argument(s).

+
+ +
[Package SparkR version 2.1.1 Index]
+ --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org For additional commands, e-mail: commits-help@spark.apache.org