|
1 | 1 | Package: analysisPipelines
|
2 | 2 | Type: Package
|
3 |
| -Title: Put analyis reports and pipelines into production |
| 3 | +Title: Compose interoperable analysis pipelines, and put them into production |
4 | 4 | Version: 0.1.0
|
5 |
| -Authors@R:c( |
| 5 | +Authors@R: c( |
6 | 6 | person("Naren","Srinivasan", email = "Naren.Srinivasan@mu-sigma.com", role = c("cre","aut")),
|
7 |
| - person("Neeratyoy","Mallik", email = "Neeratyoy.Mallik@mu-sigma.com", role = c("aut")), |
8 |
| - person("Sanjay","", email = "Sanjay@mu-sigma.com", role = c("aut")), |
9 |
| - person("Anoop S","", email = "Anoop.S@mu-sigma.com", role = c("ctb")), |
10 |
| - person("Vedavyas C","", email = "Vedavyas.C@mu-sigma.com", role = c("ctb")) |
| 7 | + person("Sanjay","", email = "Sanjay@mu-sigma.com", role = c("ctb")), |
| 8 | + person("Neeratyoy","Mallik", email = "Neeratyoy.Mallik@mu-sigma.com", role = c("ctb")), |
| 9 | + person("Anoop S","", email = "Anoop.S@mu-sigma.com", role = c("ctb")) |
11 | 10 | )
|
12 |
| -Maintainer: Naren Srinivasan <Naren.Srinivasan@mu-sigma.com> |
13 |
| -Description: This package allows data scientists to compose and generate reports as a set of analytical operations. The sequence of generation can be stored as pipelines and reused, specifically for production systems where these tasks are run repetitively. Additionally, the package implements a form of lazy evaluation where the pipelines are run on datasets only when outputs/ reports need to be generated. The package also has functions implemented for working with Spark through SparkR for both traditional Spark jobs on Spark DataFrames, as well as Spark Structured Streaming. |
| 11 | +Description: The package aims at enabling data scientists to compose pipelines of analysis which consist of data manipulation, exploratory analysis & reporting, as well as modeling steps. It also aims to enable data scientists to use tools of their choice through an R interface, and compose interoperable pipelines between R, Spark, and Python. |
14 | 12 | Depends: R (>= 3.4.0), tibble, magrittr, data.table, pipeR, devtools
|
15 |
| -Imports: ggplot2, dplyr |
16 |
| -Suggests: plotly, knitr, rmarkdown, SparkR |
| 13 | +Imports: ggplot2, dplyr, futile.logger |
| 14 | +Suggests: plotly, knitr, rmarkdown, SparkR, parallel, visNetwork, rjson, DT |
17 | 15 | Remotes: github::cran/SparkR
|
18 | 16 | Encoding: UTF-8
|
19 | 17 | License: Apache License 2.0
|
20 | 18 | LazyLoad: yes
|
21 | 19 | LazyData: yes
|
22 | 20 | RoxygenNote: 6.0.1
|
23 | 21 | VignetteBuilder: knitr
|
| 22 | +Collate: |
| 23 | + 'analysisPipelines_package.R' |
| 24 | + 'core-functions.R' |
| 25 | + 'core-functions-batch.R' |
| 26 | + 'core-streaming-functions.R' |
| 27 | + 'r-batch-eda-utilities.R' |
| 28 | + 'spark-structured-streaming-utilities.R' |
| 29 | + 'zzz.R' |
0 commit comments