First, in this release, we made huge strides with regards to Data Science. This includes:
Apache Spark 2.2 is GA. This enhances the stability of Spark for structured streaming.
Apache Zeppelin 0.7.3 is GA. This provides support for notebooks stored in HDFS, and improves the scalability and multi-user support.
Apache Livy 0.4.0 is GA. This version supports multiple programming languages in the same Spark context.
This is a point release, so most of it is around adding updated versions of different products.
Looking at the chart above, where the dotted red line is a reference point to show where we started the year, notice how site speed improvements tend to be significant and noticeable, as they are optimization-driven. Degradations, however, can generally be of any “amount,” as they happen for various reasons. LinkedIn’s page-serving pipeline has many moving parts. We deploy code multiple times per day, operate a micro-service architecture with hundreds of services, and infrastructure upgrades are frequent. A slowdown in any of these components can cause degradations.
While large degradations can be caught using A/B testing, canary analysis, or anomaly detection, small ones tend to leak to production. Thus, performance of a page has a tendency to always degrade over time.
This led to having the centralized Performance Team focus on identifying these leaks, called “site speed regressions,” and to craft tools and processes to fix them.
It’s an interesting principle. I could see this principle work for tracking database performance degradation as well.
Citing Joe Cheng, our aim is to:
- Execute long-running code asynchronously on separate thread.
- Be able to do something with the result (if success) or error (if failure), when the task completes, back on the main R thread.
A promise object represents the eventual result of an async task. A promise is an R6 object that knows:
Whether the task is running, succeeded, or failed
The result (if succeeded) or error (if failed)
This looks pretty exciting. H/T R-Bloggers
Every argument you pass to a function is a promise until the moment R evaluates it. Consider a function g with arguments x and y. Let’s leave out one argument in the function call:
g <- function(x, y) x g(1) ##  1
R will be forgiving (lazy) until the argument y is actually needed. Until then y exists in the environment of the function call as a ‘name without a value’. Only when R needs to evaluate y a value is searched for. This means that we can pass some non-existent objects as arguments to the function g and R won’t care until the argument is needed in the functions body.
Read the whole thing. Once again, H/T R-Bloggers
It’s been over a year since we first introduced introduced the Team Data Science Process (TDSP). The data, technology and practices behind Data Science continue to evolve, and the TDSP has evolved in parallel. Over the past year, several new facets have been added, including:
The IDEAR (Interactive Data Exploration, Analysis and Reporting) framework, an open source extension to R and Python designed to standardize the process of data exploration and reporting;
Guidance for use of Spark 2.0, including an end-to-end Spark v2.0 walkthrough;
Guidance for use of in-database Python with SQL Server, including an end-to-end in-database Python tutorial;
Click through for more changes, as well as links to further resources.
There are plenty of times when you might want to copy your SQL logins (including the SID) from one server to another. Perhaps you’re running an AG and need to make sure that all users exist on all your secondaries with the correct SID, perhaps you’re migrating servers and need all the logins on your new server. Whatever the reason, there are a number of different ways in which you can do this but they usually require scripting out on one server and then running the script into another server, or of course there’s Powershell.
The below script will use Python to copy all or specified logins from one server to another, including the password and SID.
Click through for the script. The main use case for a SQL Server DBA to learn Python as a DBA scripting language would be if you run SQL on Linux—I don’t think Powershell on Linux is far enough developed to handle the full range of DBA tasks. Otherwise, I’d use Powershell and dbatools.
In this module you will learn how to use the Enlighten Stack Shuffle Custom Visual. The Enlighten Stack Shuffle is helpful when you want to display a Top N set of values. For example if you want to display your top 5 selling employees this visual can make that very easy.
This looks pretty good on a dashboard, especially if you have a top-heavy data set, where the top few items are by far the most important.
You don’t have to be a data scientist to use machine learning in SQL Server. You can use pre-trained models available for usage out of the box to do your analysis. The following example shows you how you quickly get started and do text sentiment analysis.
Before starting to use this model, you need to install it. The installation is quick and instructions for installing the model can be found here: How to install the models on SQL Server
Once you have SQL Server installed with Machine Learning Services, enabled external script execution, and installed the pre-trained model, you can execute the following script to create a stored procedure that uses Python and the microsoftml function get_sentiment with the pre-trained model to determine the probability of positive sentiment of a text:
Click through to read the whole thing.
Rohan Kumar announces a new management interface for SQL Server (among other things):
These new cross-platform capabilities have made SQL Server accessible to users of Windows, Linux and Mac. At PASS Summit, we are excited to provide a sneak peek at Microsoft SQL Operations Studio. In a few weeks, users will be able to download and evaluate this free, light-weight tool for modern database development and operations on Windows, Mac or Linux machines for SQL Server, Azure SQL Database, and Azure SQL Data Warehouse. Increase your productivity with smart T-SQL code snippets and customizable dashboards to monitor and quickly detect performance bottlenecks in your SQL databases on-premises or in Azure. You’ll be able to leverage your favorite command line tools like Bash, PowerShell, sqlcmd, bcp and ssh in the Integrated Terminal window. Users can contribute directly to SQL Operations Studio via pull requests from the GitHub repository.
This appears to be a new cross platform (Mac, Linux, and Windows) for running queries against SQL Server.
Perhaps this is a replacement for SQL Server Management Studio, or may just a subset of what SSMS does today, however it works cross platform, unlike SSMS.
You won’t be giving up Management Studio anytime soon, but there are some really cool parts of SQL Operations Studio forthcoming.