Periscope Data CTO
As the complexity and volume of data grows, data teams are working quietly behind the scenes to improve their methods of analysis.
There isn’t an abundance of external conversation about the languages used to analyze these massive quantities of data, but we’ve seen first-hand here at Periscope Data that our customers are getting more sophisticated as their field matures, evolving to work with bigger data sets and integrating new techniques into their workflow. This means more than just finding new ways to innovate with SQL — advanced languages like R and Python have become a more critical part of their everyday analysis.
In this presentation, Tom O’Neill will share the impact that data analysts can see from bringing the workflows of SQL, Python and R together on one platform. Python has been one of the most popular languages for development, but for those in data environments the current workflows with Python are entirely disjointed. Python development environments don’t typically run SQL on databases, which means you’d have to query the data elsewhere, bring that data into the environment where Python runs, then take the output and migrate it back to a place where others can consume the results.
There are some who may have become accustomed to conducting R or Python analysis by connecting SQL to a Jupyter Notebook, or via an API, but we believe those solutions aren’t outcome focused. Using a notebook can make it really difficult to connect analysis back to your original data set, to keep data consistently fresh, or to get reviews from your colleagues and share results across the organization.
Analysts need functionality that enables them to add a finishing touch, be it clustering, regressions or predictions, that delivers exactly what their stakeholders need to make important business decisions. These tasks are much simpler to pull off in R or Python, and come with a whole new set of customization options for visualization.