Skip to content

Commit

Permalink
Changes for 0.3.4 release
Browse files Browse the repository at this point in the history
This PR prepares the release for 0.3.4.

This will include the changes below:

- Produces correct order of columns for nested rows when user specifies a schema 527b976
- No value in nested struct causes arrayIndexOutOfBounds (19eb277)
- `compression` aslias for `codec` option #145
- Remove dead codes, #144
- Fix nested element with name of parent bug, #161
- Minor documentation changes - #159 and #143
- Ignore comments even when it is surrounded white spaces #166

Author: hyukjinkwon <gurwls223@gmail.com>

Closes #146 from HyukjinKwon/version-0.3.4.
  • Loading branch information
HyukjinKwon committed Sep 10, 2016
1 parent 8ec82fb commit 02435b6
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 7 deletions.
12 changes: 6 additions & 6 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,26 +20,26 @@ You can link against this library in your program at the following coordinates:
```
groupId: com.databricks
artifactId: spark-xml_2.10
version: 0.3.3
version: 0.3.4
```
### Scala 2.11
```
groupId: com.databricks
artifactId: spark-xml_2.11
version: 0.3.3
version: 0.3.4
```

## Using with Spark shell
This package can be added to Spark using the `--packages` command line option. For example, to include it when starting the spark shell:

### Spark compiled with Scala 2.10
```
$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.10:0.3.3
$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.10:0.3.4
```

### Spark compiled with Scala 2.11
```
$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.11:0.3.3
$SPARK_HOME/bin/spark-shell --packages com.databricks:spark-xml_2.11:0.3.4
```

## Features
Expand Down Expand Up @@ -436,7 +436,7 @@ Automatically infer schema (data types)
```R
library(SparkR)

Sys.setenv('SPARKR_SUBMIT_ARGS'='"--packages" "com.databricks:spark-xml_2.10:0.3.3" "sparkr-shell"')
Sys.setenv('SPARKR_SUBMIT_ARGS'='"--packages" "com.databricks:spark-xml_2.10:0.3.4" "sparkr-shell"')
sqlContext <- sparkRSQL.init(sc)

df <- read.df(sqlContext, "books.xml", source = "com.databricks.spark.xml", rowTag = "book")
Expand All @@ -449,7 +449,7 @@ You can manually specify schema:
```R
library(SparkR)

Sys.setenv('SPARKR_SUBMIT_ARGS'='"--packages" "com.databricks:spark-csv_2.10:0.3.3" "sparkr-shell"')
Sys.setenv('SPARKR_SUBMIT_ARGS'='"--packages" "com.databricks:spark-csv_2.10:0.3.4" "sparkr-shell"')
sqlContext <- sparkRSQL.init(sc)
customSchema <- structType(
structField("@id", "string"),
Expand Down
2 changes: 1 addition & 1 deletion build.sbt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
name := "spark-xml"

version := "0.3.3"
version := "0.3.4"

organization := "com.databricks"

Expand Down

0 comments on commit 02435b6

Please sign in to comment.