DSS 2.1 Relase notes¶
- Migration notes
- Version 2.1.4 - October 29th 2015
- Version 2.1.3 - October 20th 2015
- Version 2.1.2 - October 6th 2015
- Version 2.1.1 - October 2nd 2015
- Version 2.1.0 - September 29th 2015
Migration to DSS 2.1 from DSS 1.X is not officially supported. You should first migrate to the latest 2.0.X version. See DSS 2.0 Relase notes
Automatic migration from Data Science Studio 2.0.X is supported, with the following restrictions and warnings:
- The “Auto Mirrors” feature, which was deprecated since DSS 1.3 has been removed
- In the Webapp builder, the deprecated API
dataiku_dataset_objecthave been removed
- If you had any webapps that did not authorize any dataset, the API key associated to this webapp is not usable anymore. You will need to create a new webapp if you want to authorize datasets.
- If you previously used Impala (in notebooks or charts), you will need to reconfigure Impala connectivity. See DSS and Impala
- Following an upgrade of the IPython notebook, graphing features must be enabled manually in older Python notebooks.
- After upgrade, you will need to follow again the procedure to Install R integration
In DSS 1.X and 2.0, DSS used an older version of the IPython notebook that had graphing features “magically enabled”. DSS 2.1 includes a more recent version of the Jupyter notebook. This new version requires manual activation of the graphing features in the notebook.
Notebooks created directly in DSS 2.1 come with a snippet that automatically enables them.
To enable them on migrated notebooks, simply add a cell at the top of your notebook containing:
DSS 2.1.4 is a bugfix release.
- Fix UI for partitioned ElasticSearch datasets
- Fix possible issue in writing to partitioned ElasticSearch datasets
- Improve handling of datasets being written while being used: Now, if a dataset is being written while being used as input of a recipe, DSS will properly remember the status of the source dataset at the beginning of the recipe, and will thus properly retrigger build if the dataset was modified during the previous recipe execution
- Fix UI for the “Time range” dependency when the output is not partitioned
- Improve the results of partitioning tests
DSS 2.1.3 is a bugfix release.
- Add support for Amazon Linux 2016.x and Ubuntu 15.10
- Improve installation R on Redhat / CentOS
- Make installation of R integration on Mac OS X more robust
- Fix publication of Jupyter notebooks to dashboard
- Fix loading of non-Filesystem datasets (like SQL tables)
- Fix loading of Parquet files with complex types (arrays, maps, structs)
- Fix scoring recipes with Random Forest algorithm
- Fix processing of charts on HDFS when Impala is not installed
- Fix aggregation on filtered grid geo charts
- Fix display of records on grid geo charts
- Fix handling of “No value” on grid geo charts
- Fix display of filtered record count on scatter plots
DSS 2.1.2 is a bugfix release.
- Code recipes can now have a partitioned dataset as input and a folder as output
- Fix handling of column types in the Split recipe > “Dispatch values of a column”
- fix failure reading indices with very large number of shards
- fix handling of sampling.
- close scroll handles on the server as soon as possible, avoids excessive resource consumption on the server
DSS 2.1.1 is a bugfix release.
For migration from previous versions, please see the DSS 2.1.0 release notes
- Fix support for Debian 7 in deps checker
- Add support for Amazon Linux 2015.09
- Fix migration of Webapp API keys in some cases
- Fix migration of ColumnRenamer processor on analysis
- Fix R automatic installation support for Debian 8
- Fix occasional failure of prepare recipes on Hadoop when starting many preparation recipes at once.
DSS 2.1.0 is a major upgrade that brings a wealth of new features and improvements.
For a summary of the major new features, see: https://learn.dataiku.com/whatsnew
DSS now features full integration with Apache Spark, the next-generation distributed analytic framework.
The integration of Spark in DSS 2.1 is pervasive and extends to all of the following features, which are now Spark-enabled:
- Visual data preparation
- “VisualSQL” recipes (Grouping, Joining, Stacking)
- Guided machine learning in analysis
- Training and prediction in Flow
- PySpark recipe
- SparkR recipe
- SparkSQL recipe
- PySpark-enabled notebook
- SparkR-enabled notebook
All DSS data sources can be handled using Spark. As always in DSS, you can mix technologies freely, both Spark-enabled and traditional
For more information about Spark integration, see DSS and Spark
Plugins let you extend the features of DSS. You can add new kinds of datasets, recipes, visual preparation processors, custom formula functions, and more.
Plugins can be downloaded from the official Dataiku community site, or created by you and shared with your team.
DSS now includes a REST API to programmatically manage DSS from any HTTP-capable language.
To learn more about this API, see Public REST API
The charts module of DSS has been vastly enhanced:
New chart types
- Horizontal bar charts
- Pie and donut charts
- Pivot table (text view) and colored pivot table
- Geographic scatter map
- “Grid” map (fixed-width aggregation grid)
- (Experimental) 2D distribution plot
Brand new user experience. It is now much easier to understand what’s going in your chart, and to switch dimensions and measures.
Improved presentation of date axis
All charts can now have semi-transparency
New computation mode for aggregated charts: “Percentage scale” (% of the total)
Tooltips now provide the ability to drill-down and exclude values
New R integration and Jupyter notebook¶
The bundled IPython notebook has been upgraded. DSS now includes Jupyter 4.0
This is a major new release of IPython/Jupyter, with lots of improvements, and the support for multiple languages.
The bundled Jupyter now comes builtin with a brand new R kernel, with vastly improved features over the previous R integration:
- Syntax highlighting
- Auto-completion (just hit Tab)
- Much improved error handling
In addition, DSS features a new R API. See R API for more details.
Editable datasets are a new kind of dataset in DSS, which you can directly create and modify in the DSS UI, ala Excel or Google Spreadsheets.
They can be used for example to create referentials, configuration datasets, …
Editable datasets can be imported from any file, or another dataset.
DSS comes with a large number of supported formats, machine learning engines, … But sometimes you want to do more.
DSS code recipes (Python and R) can now read and write from “Managed Folders”, handles on filesystem-hosted folders, where you can store any kind of data.
DSS will not try to read or write structured data from managed folders, but they appear in Flow and can be used for dependencies computation. Furthermore, you can upload and download files from the managed folder using the DSS API.
Here are some example use cases:
- You have some files that DSS cannot read, but you have a Python library which can read them: upload the files to a manged folder, and use your Python library in a Python recipe to create a regular dataset
- You want to use Vowpal Wabbit to create a model. DSS does not have full-fledged integration in VW. Make a first Python recipe that has a managed folder as output, and write the saved VW model in it. Write another recipe that reads from the same managed folder to make a prediction recipe
- Anything else you might think of. Thanks to managed folders, DSS can help you even when it does not know about your data.
Previously, DSS could use Impala:
- For charts creation
- In the Impala notebook
You can now also use Impala in a new Impala recipe to benefit from the speed of this engine for aggregations on HDFS.
Impala is also available as an engine for “VisualSQL” (Grouping, Join, Stack) recipes
There are times when you just need to:
- Run a shell command
- Stream a dataset through a shell command (stdin/stdout)
The shell recipe lets you do just that.
In all modules of DSS where you can write code, you now have the ability to insert code snippets. DSS comes builtin with lots of useful snippets, and you can also write your own and share them with your team.
DSS now has full read-write support for ElasticSearch datasets
Will you find all our new easter eggs ?
Multiple SQL statements support¶
Every module of DSS where you can write SQL code now support multiple statements.
For example, in the SQL Query recipe, you can now add statements before the main “SELECT” statement. This can allow you to issue some SET statements to tune the optimizer, or to create stored procedures.
In other words, you can now declare and call a stored procedure in a SQL Query recipe.
This also applies for the pre-write and post-write statements. For example, you can now create multiple indexes.
Running SQL, Hive and Impala from Python and R recipes¶
SQL is the most pervasive way to make data analysis queries. However, doing advanced logic, like loops, conditions, … is often difficult in SQL. There are some options like stored queries, but they require learning new languages.
DSS now lets you run SQL queries directly from a Python recipe. This lets you:
- sequence several SQL queries
- dynamically generate some new SQL queries to execute in the database
- use SQL to obtain some aggregated data for further numerical processing in Python
To learn more, see API for performing SQL, Hive and Impala queries
Thanks to “repartitioning” mode, you can now:
- start with a non-partitioned files-based dataset where a column could act as a partitioning column
- create a “repartitioning-enabled” sync or prepare recipe to transform it to a partitioned dataset
- Build in a single pass all partitions of the target dataset
Several recipes and datasets now support “Append” instead of “Overwrite” in target datasets.
While we do not recommend to use this for general-purpose datasets, it can be useful in some cases, like for example create a “history” dataset as output of a recipe, writing a new line each time the recipe is run.
HDP 2.3 support¶
DSS is now compatible with Hadoop HDP 2.3.0
- Added detection of US states in Preparation recipes
- New processor to concatenate arrays
- The column renamer processor can now perform multiple renames
- New mass actions to lowercase/uppercase/simplify many column names
- Find/Replace processor can now work on all columns at once
- Analysis: Machine Learning training does not start automatically anymore, giving you an opportunity to review the settings prior to initial training.
Custom formula: new random() function
Automatic R integration installation, no more manual steps
- All recipes now share a more common and more consistent layout, with the Run button always present on the most important tab.
- When creating managed datasets, you now have more options to preconfigure the dataset based on common formats, or using known partitioning schemes
- Better build dataset modal, with more clear explanation of the build modes
- Fixed various scrolling and drag-and-drop issues
- Fix renaming of datasets when the name was already used
- Fix project home display for readers
- Fixed job preview when training a model
- It’s now possible to choose a different SQL port
- Improved behavior of exports
- Fixed tagging of analysis
- Fixed DSS disconnection when very long log lines are transferred
- Fixed an issue with date range filters
- Fixed a bug that could happen with numerical axis and small values
- Fixed issue in geo chart when the value of the aggregation is 0
- Fixed various issues in legend, especially with “Mixed columns/lines” charts
- [HDP only] Fixed Hive recipes on readonly HDFS datasets
- Fixed/Added support for native Snappy and LZO datasets
- Fixed EXPLAIN statements in notebook for Hive
- The command-line tool to import Hive databases now supports partitioning
SQL and VisualSQL recipes¶
- Added ability to disable the “LIMIT” in statements for some operations that don’t support it (e.g. tuple mover operations on Vertica)
- Fixed cases where the “limit” statement could get disabled
- Fixed issues with Join recipe on Oracle
- Grouping recipe: fixed issue with “LAST” aggregations function
- Fixed typing of MIN/MAX on non-string columns
- Grouping recipe can now use faster stream engine when computing COUNT
- Fixed Gradient Boostring Tree in kfold mode with multiple losses
- Fixed Lasso regression in AutoCV mode
- SGD classification can now use sparse matrixes
- Fixed error when imputation was disabled for all numerical features
- Fixed error when using feature hashing on a numerical feature
- Fixed broken models when RSMLE can’t be computed
- Fixed some errors in notebook export
- Using “Compute number of records” on large datasets does not lock the studio anymore
- Fixed a deadlock leading to the whole notification system ceasing to work
- Fixed the infamous “error during execution of add command” issue with the internal Git repository
- Fixed several cases where long operations could be performed while holding a lock on the DSS configuration (like Hive validation or aborting Impala queries)
- DSS diagnosis now includes more useful information
- Fixed detection of Excel dates
- Fixed detection of UTF-8 BOMs
- Shapefiles that require Vecmath now display a proper informative message
- We now display proper warning messages when using non-Hive-compatible HDFS dataset names