Data is one of those things that’s crucial for any company, but also something that no one likes to deal with.
Well, that era is over.
Why?
Not only is data getting more and more plentiful, but analyzing and managing it well is getting increasingly crucial for any business. As we’ve discussed here) As a result, the role of business analysts is h5 significance.
At the same time, tools like ChatGPT (and especially the new code interpreter) greatly enhance data analysts’ capabilities.
So today, we’re going to dive into why analysts ought to embrace, and not fear AI, given how it augments, rather than supplants them.
Instead of discussing hypotheticals, we’re going to discuss tools that exist today, and how they can help your day to day workflow as an analyst, using a play by play framework.
We’ll have a look at an analyst’s daily workflow, see what inefficiencies these tasks within it may conventionally come with, and explain how these inefficiencies can be solved, using a practical lens. Furthermore, we’ll also have a look at new capabilities that analysts of different skill sets can leverage using AI tools h5 popularity today.
A business analyst's day often commences with parsing through a multitude of emails alerting them to fresh batches of data to be analyzed. This could be transaction data, usage reports from APIs, or spreadsheets on customer feedback.
Manual data collection, though integral, tends to be quite inefficient. It involves individually connecting to each source, which ranges from executing custom scripts for databases to making API calls and manually extracting data from spreadsheets. These tasks can become particularly cumbersome when considering technical challenges such as occasional server downtime, access restrictions, or inconsistent data formatting across sources.
Upon successful data collection, analysts must validate the data's integrity, a step necessary for ensuring the accuracy of any subsequent analysis. This process involves checking for data anomalies, identifying missing values, and verifying data consistency.
Evidently, there’s a lot of tasks within any analysts workflow that, if made more efficient, would be a boon for the analysts and their teams alike.
Fortunately, though, we’re living in an era when there’s game-changing tools coming out seemingly every week-provided you actually stay plugged in, of course (but that’s what we’re for!)
So let’s have a look at some of the best tools you can use to make the aforementioned aspects of your workflow as a data analyst easier:
Levity.AI can significantly streamline and enhance the process of parsing through emails. Given how their data pipelines are particularly crucial for analysts and how crucial daily emails are for that, this tool can prove to be invaluable:
By integrating Levity into an analyst’s workflow, the initial phase of their day - sifting through their data pipeline- becomes drastically easier to handle. They can begin their day with an organized and prioritized list of data tasks, enabling them to focus more on actually deriving actionable insights from the data.
If you’re reading this, I have to assume you know about ChatGPT. You may not know about the new Code Interpreter, or at least not know about how it can tangibly make your life as an analyst easier.
Let’s have a look!
Suppose you’re working with a relational database like MySQL, PostgreSQL, or SQLite. You’d typically need to manually connect to the database, write a SQL query to select the relevant data, execute the query, and then export the results for further analysis.
With the new ChatGPT Code Interpreter, you can write a Python script that automates this process.
Showing tends to be better than telling, so I figured i’d just go to the horse’s mouth and asked GPT4+Code Interpreter how it’d go about this. Here’s its answer:
Now a hypothetical framework is one thing.
How would using the tool look like in your everyday workflow as an analyst?
Churn is obviously an incredibly important metric for any product. So, being able to accurately predict which users are more likely to is as well.
Let’s see how the new Code Interpreter can help with that:
Traditionally, predicting user churn at a place like Netflix or Spotify involves a series of fragmented steps.
Analysts might start by using SQL to extract data from a database, then switch to Python or R for cleaning the data and engineering features, and finally use a machine learning library or statistical software like SAS or SPSS to build and tune the predictive model. This disjointed process can be time-consuming and prone to errors due to the need to switch between different tools and languages.
Comparatively, the Code Interpreter can guide analysts through the entire process within a single, integrated Python-based workflow.
Using data from Kaggle’s ‘Telco Customer Churn’ dataset (hypothetical data about a telecommunications company), let’s see how the Code Interpreter can create a model to predict the likelihood of individual customer churn rates.
1: Code Interpreter identifies the various attributes within the data that are relevant at predicting churn:
2: Code Interpreter autonomously spots a discrepancy between how data is presented, and tweaks it to make it workable:
For the sake of brevity, I haven’t here, but if you click on ‘show work’, you can literally see the code the model has employed to accomplish this.
3: Code Interpreter demonstrates its ‘thought process’ behind which method of removal it opts for:
4: Code Interpreter performs Exploratory Data Analysis
The following is the ‘work’ the model used to accomplish the same:
5: Code Interpreter employs feature engineering
6: Code Interpreter builds its predictive model
7: Code Interpreter evaluates the viability of its own model:
Obviously, that’s quite a bit to go through. The thing is, it’s so intuitive that honestly I figured I might as well just show and not tell.
Here’s the TLDR:
This entire process essentially resulted in a Logistic Regression model that can predict customer churn based on various features.
In addition to obviously automating several elements of an analysts' workflow, this sort of functionality (that is the worst it'll ever be) also drastically lowers the technical barrier for being an effective analyst as well.
As the Code Interpreter shows, AI can be massively helpful with analysts' workflows. The trick is with regards to how it's implemented; in a nutshell, the UX.
Crunch leverages AI and a nifty UX made possible by it to create what is essentially a co-pilot for analytics.
Want to see what it's like?
Signup for our waitlist here!