Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

This document details steps to setup new applications using Sprngy Admin UI, defining models and running workloads.

Table of Contents
minLevel1
maxLevel3
Demo Case
outlinefalse
stylenone
typelist
printablefalse

Use Case Category: Finance

Example

Demo Application 1: Assess the Risk of a Loan

Objective - Determine whether a loan will be approved or not by the issuing bank

Overview

From source:

“The bank wants to improve their services. For instance, the bank managers have only vague idea, who is a good client (whom to offer some additional services) and who is a bad client (whom to watch carefully to minimize the bank loses). Fortunately, the bank stores data about their clients, the accounts (transactions within several months), the loans already granted, the credit cards issued The bank managers hope to improve their understanding of customers and seek specific actions to improve services. A mere application of a discovery tool will not be convincing for them.

The data about the clients and their accounts consist of following relations:

  • relation account - each record describes static characteristics of an account,

  • relation client - each record describes characteristics of a client,

  • relation disposition - each record relates together a client with an account i.e. this relation describes the rights of clients to operate accounts,

  • relation permanent order - each record describes characteristics of a payment order,

  • relation transaction - each record describes one transaction on an account,

  • relation loan - each record describes a loan granted for a given account,

  • relation credit card - each record describes a credit card issued to an account,

  • relation demographic data - each record describes demographic characteristics of a district.

Each account has both static characteristics (e.g. date of creation, address of the branch) given in relation "account" and dynamic characteristics (e.g. payments debited or credited, balances) given in relations "permanent order" and "transaction". Relation "client" describes characteristics of persons who can manipulate with the accounts. One client can have more accounts, more clients can manipulate with single account; clients and accounts are related together in relation "disposition". Relations "loan" and "credit card" describe some services which the bank offers to its clients; more credit cards can be issued to an account, at most one loan can be granted for an account. Relation "demographic data" gives some publicly available information about the districts (e.g. the unemployment rate); additional information about the clients can be deduced from this.”

Classification:

The file structure is going to look like this:

Download the files:

Districtdistrict

Accountaccount

Clientclient

Loanloan

Orderorder

Transactiontransaction

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Data Warehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Step 1: Setting up the Application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application Configuration tab, select Create New, and filling out the file structure. Make sure to indicate District as the parent of Account and Client, Account as the parent of Loan, Order, Transaction, Display, and Display as the parent of Card.

We will begin first with the District Dataset:

Step 2: Meta Model, Ingest Model, and Workloads

We are going to use the Bulk Upload Meta feature to create our meta model for the District dataset. In the AdminUI menu , under the Meta Model dropdown, select the Bulk Upload Meta option. Select the Loan Application module and District entity, and import the meta model CSV file into the bulk upload. Once you hit submit, this will automatically generate the meta model for the District dataset.

Download the csv files for meta and ingest model:

Account Meta Model

Account Ingest Model

Card Meta Model

Card Ingest Model

Client Meta Model

Client Ingest Model

Display Meta Model

Display Ingest Model

District Meta Model

District Ingest Model

Loan Meta Model

Loan Ingest Model

Image Removed

Next, similar to meta model you can bulk upload ingest model in AdminUI. Or create ingest model by selecting processors. You can refer to this page to understand what each of the processors do.

Image Removed

Once we submit the Ingest Model, we can Image Added

Next, similar to meta model you can bulk upload ingest model in AdminUI. Or create ingest model by selecting options.

Image Added

Once we submit the Ingest Model, we can run the workloads under the Workload Management/Run Workloads page.

Once you confirm in HDFS that the SDL-FDL workload ran correctly, run the FDL-BDL workload next. This will apply the processors we selected in the Ingest Model for the FDL to BDL layer. You can see if the workload ran correctly by going to the LOANAPPLICATION/District/BDL/Fact directory in HDFS

Data Upload

Upload data using 'Data Upload' option. Select the Loan Application module and District entity, and CSV file into the data upload. Once you hit submit, this will automatically upload the district dataset to SDL/Land folder in datalake.

SpringyBi

Visualizations

After adding the datasets, you can select any one of them to bring up the Charts page. Here, you can pick which type of visualization you would like to create and the parameters for that visualization. Below are three examples of charts you can make in Superset SprngyBI that show the correlations between the various districts, regional statistics, loan amounts and durations, and the status of the loans.

 


Demo

Use Case Category: Sports

Example

Demo Application 1: Evaluating NBA players' performances relative to their salaries

Objective - Build the most efficient team while constraining to a team’s cap space

Overview

With the NBA’s increased emphasis on data analytics, team managers have access to numerous metrics that show a player’s offensive and defensive efficiency. One such metric is FiveThirtyEight’s RAPTOR method, which highlights a player’s net efficiency and wins above replacement (WAR). In 2022, the top five players in Total RAPTOR were as following:

Player

Total RAPTOR

2022 Salary (in $M)

Nikola Jokic

+ 14.6

$31.5

Giannis Antetokounmpo

+ 8.1

$39.3

Joel Embiid

+ 7.8

$31.5

Rudy Gobert

+ 6.9

$35.3

Stephen Curry

+ 6.4

$45.8

While it would be a manager’s dream to have a starting lineup like this, their combined salaries exceeds the cap space of $122M, leaving the team with both a hefty luxury tax bill and no bench players. The goal of this application is to build a full NBA team that gets the most value per dollar spent, according to the RAPTOR metric.

Classification:

Before defining the application in the system, it is useful to identify key entities and classify the application.

Key Entities:

The file structure is going to look like this:

Download the files:

team

players

salaries

stats

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Data Warehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Explanation of the classification: <>

Normalization: <V Structure/Star Model/Snowflake and its explanation>

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tabConfiguration page, select Create New, and filling out the file structure. Make sure to indicate Team as the parent of Players, and Players as the parent of Salaries and Stats.

We will begin first with the Teams dataset.

Data Upload

For uploading a csv file of teams dataset go to Data Upload from left side meu. Select Module name as NBASTATISTICS and entity name as teams from the dropdown option. Select the file to be uploaded and click on Submit.

On submit the dataset gets uploaded to datalake for that entity.

Step 2: Meta Model, Ingest Model, and Workloads

We can now set up the Meta Model in AdminUI:

Image RemovedNext, create the ingest model in AdminUI. The first part will be defining which processors to use from SDL to FDL. These are the SDL-FDL processors we select for our Teams data. You can refer to this page to understand what each of the processors doImage Added

Next, create the ingest model in AdminUI.

Once we submit the Ingest Model, we can run the workloads under the Batch Management/Run Workloads page.

The Ingest Model will be similar to the one we made for the team’s dataset, the only difference being we define two extra processors for the nested quality of the players datasetnesting will be required to be turned on.

After creating the Ingest Model, run the SDL-FDL and FDL-BDL workloads, checking the respective directories to make sure that both workloads ran correctly.

Repeat the same steps for rest of the entities.

SpringyBI

Visualizations

Step 4: Importing Dataset into

Superset

SprngyBI

Once you are in SupersetSprngyBI, select the Datasets option from the Data dropdown in the top menu. From there, select the add Dataset option. Set the Database to Apache Hive, select the NBAstatistics database from Schema, and select which table you would like to add. Superset SprngyBI will only allow you to add one table at a time, but you can add as many tables as you want one by one.

(see this page for further reference)

After adding the datasets, you can select any one of them to bring up the Charts page. Here, you can pick which type of visualization you would like to create and the parameters for that visualization. Below are two examples of charts you can make in Superset SprngyBI that show individual performance relative to salary and each team’s total offensive and defensive efficiency.

Example

Demo Application 2

-

: Data Analysis of IPL

2022

2022 (Indian Premier League) cricket match dataset.

Objective: Analyzing the IPL 2022 match by match dataset.

Overview: This is the IPL 2022 match by match dataset which includes information like match details, team details, winning team name, winning margin, player of the match etc.

IPL dataset:

View file
nameIPL_Matches_2022.csv

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

 ✔

Modelling

Info

This involves building statistical models and testing those.

 

 

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

 ✔

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

 

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

 ✔

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tab, select Create New, and filling out the file structure.

Step 2 : Data Upload

We will begin with the Match dataset.

Go to the Data Upload page from the side navigation. Select module name, entity name and the csv file.

Step 2: Meta Model, Ingest Model, and Workloads

We can now set up the Meta Model in AdminUI:

This is how our meta model looks after adding all the columns with their respective datatypes. Note that you do not add the as_of_date column, as that will be added automatically.

Next, create the ingest model in AdminUI. The first part will be defining which processors to use from SDL to FDL. These are the SDL-FDL processors we select for our Teams data. You can refer to this page to understand what each of the processors do.

Once we submit the Ingest Model, we can run the workloads under the Batch Management/Run Workloads page.

First, run the SDL-FDL workload. This will apply the processors we selected in the Ingest Model for the SDL to FDL layer. You can see if the workload ran correctly by going to the IPL/Match/FDL/Stage directory in HDFSFDL workload.

Once you confirm in HDFS that the SDL-FDL workload ran correctly, run the FDL-BDL workload next. This will apply the processors we selected in the Ingest Model for the FDL to BDL layer. You can see if the workload ran correctly by going to the IPL/Match/BDL/Fact directory in HDFS.

Step 4: Importing Dataset into

Superset

SprngyBI

Once you are in SupersetSprngyBI, select the Datasets option from the Data dropdown in the top menu. From there, select the add Dataset option. Set the Database to Apache Hive, select the IPL database from Schema, and select which table you would like to add. Superset SprngyBI will only allow you to add one table at a time, but you can add as many tables as you want one by one.

(see this page for further reference)

After adding the datasets, you can select any one of them to bring up the Charts page. Here, you can pick which type of visualization you would like to create and the parameters for that visualization.

Below is the example of number matches won by a team in IPL 2022.

Demo


Use Case Category: Retail

Example

Demo Application 1

-

: Exploratory Data Analysis of Car Sales dataset.

Objective: Car Sales dataset will give abstracts like:

  • Percentage of car sales done by the car companies.

  • Correlation of the car sales based on engine size and fuel capacity.

  • Power performance of cars for each manufacturers based on the relationship between Horsepower and the length of the cars.

  • Visualization of the number of cars manufactured and the horsepower with what they are manufactured

  • The car companies that did the most sales based on the average horsepower of all their models and average price for all those models .

Overview: This is the Car sales data set from the years 2008 to 2012, which include includes information about different cars . Here, we have to see which features have more impact on car sales depending on the different attributes given.

Car Sales Dataset:

View file
nameCar_sales.csv

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tabConfiguration screen, select Create New, and filling out the file structure. Since we have just one layer file system data we will have just one entity in it.

Data Upload

Go to data upload page from side navigation. Select module name, entity name and csv of the data file. This will upload data to datalake.

Step 2: Meta Model, Ingest Model, and Workloads

We can now set up the Meta Model in AdminUI:

Add the column names and data types from the teams dataset into the Create Meta Model page and then click submit. Note that you do not add the as_of_date column, as that will be added automatically.

Next, create the ingest model in AdminUI. The first part will be defining which processors to use from SDL to FDL. These are the SDL-FDL processors we select for our Teams data. You can refer to this page to understand what each of the processors do.

Once we submit the Ingest Model, we can run the workloads under the Batch Management/Run Workloads page.

First, run the SDL-FDL workload. This will apply the processors we selected in the Ingest Model for the SDL to FDL layer. You can see if the workload ran correctly by going to the FDL/Stage directory in HDFSFDL workload.

Once you confirm in HDFS that the SDL-FDL workload ran correctly, run the FDL-BDL workload next. This will apply the processors we selected in the Ingest Model for the FDL to BDL layer. You can see if the workload ran correctly by going to the BDL/Fact directory in HDFS.

Step 4: Importing Database and Dataset into

Superset

SprngyBI

After adding the database and datasets (as shown in above screenshots) click on the dataset that you have created to bring up the Charts page. Here, you can pick which type of visualization you would like to create and the parameters for that visualization. Next step is to create a new dashboard for displaying all the charts at one place. Below are the examples of the charts you can make in SupersetSprngyBI, that shows analysis like:

  • Percentage of car sales done by the car companies.

  • Correlation of the car sales based on engine size and fuel capacity.

  • Power performance of cars for each manufacturers based on the relationship between Horsepower and the length of the cars.

  • Visualization of the number of cars manufactured and the horsepower with what they are manufactured

  • The car companies that did the most sales based on the average horsepower of all their models and average price for all those models .

Example

Demo Application 2

-

: Exploratory Data Analysis of eCommerce dataset.

Objective: Analysis of eCommerce dataset will give the information about:

  • Yearly spent by every individual on individual brands.

  • Yearly spent based on the number of membership years.

  • Number if items users bought.

  • Majorly sold products on eCommerce platform.

  • Comparison between time spent by users on app vs on website.

  • Analysis of which category of products have the highest price.

Overview: eCommerce Dataset has two entities: Customers and Sales. Customers being the parent and Sales being the child entities. Sales dataset has all the details about what all products the user has bought and the Customers dataset has all the details about the user’s profile.

View file
nameSales.csv
View file
nameCustomers.csv

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tabConfiguration page, select Create New, and filling out the file structure. ce weSince Since we have just one layer file system data we will have just one entity in it.

Step 2: Upload the dataset to the Land directory of the given entity.

Go to data upload page from side navigation. Select module name, entity name and csv of the data file. This will upload data to datalake.

Step 3: Creating Bulk Upload Meta Model

Bulk uploads is an option given in the UI for a user who wants to upload a CSV file of their Meta, Ingest, Analytic, or Import Models. Once uploaded, the respective models will get created in the HDFS.

Create the bulk upload meta model for customers and sales using below attached csv files through UI navigating to Meta Model → Bulk Upload Meta.

View file
nameecommerce_sales_meta_modelSales.csv
View file
nameecommercemeta_customers_meta_model.csv

Image Modified

Step 4: Creating Bulk Upload Ingest Model

Bulk uploads is an option given in the UI for a user who wants to upload a CSV file of their Meta, Ingest, Analytic, or Import Models. Once uploaded, the respective models will get created in the HDFS.

Create the bulk upload ingest model for customers and sales by uploading below attached csv files through UI navigating to Ingest Model → Bulk Upload Ingest.

View file
nameecommerce_sales_ingest_modelSales.csv
View file
nameecommerceingest_customers_ingest_model.csv

Step 5: Running the Workloads

First, run the SDL-FDL workload. This will apply the processors we selected in the Ingest Model for the SDL to FDL layer. You can see if the workload ran correctly by going to the FDL/Stage directory in HDFS.

Once you confirm in HDFS that the SDL-FDL workload ran correctly, run the FDL-BDL workload next. This will apply the processors we selected in the Ingest Model for the FDL to BDL layer. You can see if the workload ran correctly by going to the BDL/Fact directory in HDFSnext.

Step 6: Importing Database and Dataset into

Superset

SprngyBI and creating a dashboard for the charts

Once you are in SupersetSprngyBI, select the Datasets option from the Data dropdown in the top menu. From there, select the add Dataset option. Set the Database to Apache Hive, select your database from Schema, and select which table you would like to add. Superset SprngyBI will only allow you to add one table at a time, but you can add as many tables as you want one by one.

(see this page for further reference)

Image Modified

Visualization of eCommerce dataset will give the information about:

  • Yearly spent by every individual on individual brands.

  • Yearly spent based on the number of membership years.

  • Number if items users bought.

  • Majorly sold products on eCommerce platform.

  • Comparison between time spent by users on app vs on website.

  • Analysis of which category of products have the highest price.

Demo

Use Case Category: Entertainment

Example

Demo Application 1

-

: Exploratory Data Analysis of a Movie Dataset (IMDB)

Objective: Given the IMDB dataset for seven different entities we will visualize:

  • The popularity of a movie and it’s genres based on the movie rank given.

  • Relationship between the movies and directors, which can also give the abstract of which all directors directed the most popular movies.

  • Details of the directors and the genres on which they make their movies.

  • Abstract of famous actors of the movies and their popularity based on the roles they played in a movie.

Overview: IMDB Dataset has total 7 different entities, which gives all the details for a particular movie like the popularity of movie, directors, actors, genre of the movies that the actors do and genre of the movies that the directors direct, below is the dataset of IMDB in sql format.

View file
nameimdb (1).sql

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tabConfiguration page, select Create New, and filling out the file structure. Since we have just one layer file system we will have just one entity in it.

Since the dataset was in SQL format, it was first imported by using the following command in terminal and then it was converted into parquet format by executing the import model and through which it was stored on SDL Land.

Code Block
languager
sudo mysql imdb < /home/bapdevvm/Downloads/imdb.sql 

Import Model:

The above is for one entity i.e., “actors“, there are total 7 entities and have to follow similar steps for other 6 entities by just changing the table names as per entities.

Entities:

Movie Genres

Movie

Actors

Roles

Movies Directors

Directors

Director Genres

The relationship between the parent-child entities is shown by below flowchart:

Step 2: Meta Model, Ingest Model, and Workloads

We can now set up the Meta Model in AdminUI:

or can use bulk upload if you have the csv file available in correct format.

Ingest Model:

Ingest model for the only parent (root) entity:

Similarly add ingest model for all other entities either by using the UI or by bulk uploading the csv.

Run Import Workloads:

One need to follow the above steps for all the other entities to run the import workload.

Importing Database and Dataset into

Superset

SprngyBI and creating a dashboard for the charts

Creating a dataset for two child entities i.e. for “actors” table and for “director_genres” table.

Click on the dataset to create charts for the visualizations. Create a dashboard to save all the charts that you have created. Below are the charts created to give the abstracts like:

  • The popularity of a movie and it’s genres based on the movie rank given.

  • Relationship between the movies and directors, which can also give the abstract of which all directors directed the most popular movies.

  • Details of the directors and the genres on which they make their movies.

  • Abstract of famous actors of the movies and their popularity based on the roles they played in a movie.

Below is the visualization of the IMDB dataset:

Demo

Use Case Category:

World

Airports

Example

Demo Application 1: Mapping World Airports

Objective - Visualize the number of airports by country and state

Overview

Source

We will plot the location of each airport that is in the airports dataset. We will also visualize which country and region/state has the greatest number of airports.

Classification:

Before defining the application in the system, it is useful to identify key entities and classify the application.

Key Entities:

The file structure is going to look like this:

Download the files:

airports

frequencies

runways

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Explanation of the classification: <>

Normalization: <V Structure/Star Model/Snowflake and its explanation>

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application tabConfiguration page, select Create New, and filling out the file structure. Make sure to indicate Airports as the parent of Frequency, and Frequency as the parent of Runways.

Step 2: Upload Data

Go to data upload page from side navigation. Select module name, entity name and csv of the data file. This will upload data to datalake.

Step 3: Meta Model, Ingest Model, and Workloads

We can now set up the Meta Model in AdminUI:

Next, create the ingest model in AdminUI. You can refer to this page to understand what each of the processors do.

Once we submit the Ingest Model, we can run the workloads under the Batch Management/Run Workloads page.

Importing Dataset into

Superset

SprngyBI

Once you are in SupersetSprngyBI, select the Datasets option from the Data dropdown in the top menu. From there, select the add Dataset option. Set the Database to Apache Hive, select the WORLDAIRPORTS database from Schema, and select which table you would like to add. Superset SprngyBI will only allow you to add one table at a time, but you can add as many tables as you want one by one.(see this page for further reference)

After adding the datasets, you can select any one of them to bring up the Charts page. Here, you can pick which type of visualization you would like to create and the parameters for that visualization. Below are two examples of charts you can make in SupersetSprngyBI:

Demo

Use Case Category: Education

Example

Demo Application 1: To assess the role of a person in the University

Objective - To get the information about the role of a person in the university:

  • Visualization of professor and students individually at each individual course level

  • Relationship between Student, the phase they are in currently and for the number of years they are in a university program

  • How many students are advised by each professor

  • How many percentage of people teaching at each course level

Overview: The dataset has got 4 different entities which gives information about the relationship and details of the person, course and the role they have in the university, below is the dataset in sql format.

View file
nameUW_std.sql

Classifying the Application:

Based on the data technology used and the business use, the application is classified as below:

Quality of Data

Data Lake

Info

Data Lake is a centralized repository to store large amount of raw data

Data Lakehouse / Data Warehouse

Info

Data Lakehouse combines concepts of Data Lake and Datawarehouse providing large storage capacity combined with data management features

Curated

Info

Curating data involves creating or preparing data to make it usable for business analysis.

Correlated

Info

Correlated data means running Algorithms to discover patterns and relationships within the data.

Normalized

Info

Normalizing data involves structuring the data to enable rapid access.

Analyze

Info

Data Analysis involves identifying useful information to support decision-making, often using visualizations

Modelling

Info

This involves building statistical models and testing those.

Step 1: Setting up the application

Now that the business use and classification of the application is established, the application can be created using the UI. In AdminUI, set up the application by going to the Set-up Application Configuration tab, select Create New, and filling out the file structure.

Since the dataset was in SQL format, it was first imported by using the following command in terminal and then it was converted into parquet format by executing the import model and through which it was stored on SDL Land.

Code Block
languager
sudo mysql imdb < /home/bapdevvm/Downloads/UW_std.sql 

Import Model:

There are total 4 entities and have to follow the same steps as done above to create an Import model for other 3 entities i.e. “course“, “person, “taughtBy“.

Here all the entities are considered as parent and there is no relationship assigned between any of these entities.

Step 2: Meta Model: We can now set up the Meta Model in AdminUI. This is just for one entity “advisedBy“ and have to do the same set up for other 3 entities.

Upload the csv for the given entity and then click submit. Note that you do not add the as_of_date column, as that will be added automatically

Step 3: Ingest Model: This is just for one entity have to do same set up for other entities individually.

Create the ingest model in AdminUI. You can refer to this page to understand what each of the processors do.

Step 4: Running Import Workloads: Batch Management → Run Import workload: This is just for one entity have to do same set up for other entities individually.

Importing Database and Dataset into

Superset

SprngyBI and creating a dashboard for the charts

Once you are in SupersetSprngyBI, select the Datasets option from the Data dropdown in the top menu. From there, select the add Dataset option. Set the Database to Apache Hive, select your database from Schema, and select which table you would like to add. Superset SprngyBI will only allow you to add one table at a time, but you can add as many tables as you want one by one.(see this page for further reference)

Below screenshot of dataset is for one entity, can create for other 3 remaining entities.

Step 10: Visualization

Below is the visualization:

Copyright © Springy Corporation. All rights reserved. Not to be reproduced or distributed without express written consent.