BI UNIT 2 & 3
BI UNIT 2 & 3
Decision making is the process of making choices by identifying a decision, gathering information,
and assessing alternative resolutions.
Using a step-by-step decision-making process can help you make more deliberate, thoughtful
decisions by organizing relevant information and defining alternatives. This approach increases the
chances that you will choose the most satisfying alternative possible.
In a JIT inventory system, the organization requires real-time data of their inventory
levels to place orders “just in time” to prevent delays in production and cause a
negative domino effect. Therefore, a DSS is more tailored to the individual or
organization making the decision than a traditional system.
The model management system S=stores models that managers can use in their
decision-making. The models are used in decision-making regarding the financial
health of the organization and forecasting demand for a good or service.
2. User Interface
The user interface includes tools that help the end-user of a DSS to navigate through
the system.
3. Knowledge Base
The knowledge base includes information from internal sources (information collected
in a transaction process system) and external sources (newspapers and online
databases).
Types of Decision Support Systems
Communication-driven: Allows companies to support tasks that require
more than one person to work on the task. It includes integrated tools such as
Microsoft SharePoint Workspace and Google Docs.
Model-driven: Allows access to and the management of financial,
organizational, and statistical models. Data is collected, and parameters are
determined using the information provided by users. The information is
created into a decision-making model to analyze situations. An example of a
model-driven DSS is Dicodess – an open-source model-driven DSS.
Knowledge-driven: Provides factual and specialized solutions to situations
using stored facts, procedures, rules, or interactive decision-making structures
like flowcharts.
Document-driven: Manages unstructured information in different electronic
formats.
Data-driven: Helps companies to store and analyze internal and external
data.
Intelligence Phase
Design Phase
Choice Phase
Implementation Phase
Intelligence Phase
Firstly, the decision-making process starts with the intelligence phase. In the
first phase, decision makers examine reality and try to identify problems or
opportunities correctly. This phase is not only related to the Simon’s decision-
making process, but also to other fields and other methodologies. For
example, we like to practice Lean Startup methodology which emphasizes the
importance of right problem definition before building anything (product or
business).
The intelligence phase can last really long. But, since decision-making process
starts with this phase, it should be to be done properly. This is a key
ingredient in every business success.
Design Phase
The main goal of the design phase is to define and construct a model which
represent a system, by defining relationships between collected variables.
Once we validate the model, we define the criteria of choice and search for
several possible solutions for the defined problem (opportunity). We wrap up
the design phase by predicting the future outcomes for each alternative.
Choice Phase
In this phase we are actually making decisions. The end product of this phase
is a decision. Decision is made by selecting and evaluating alternatives
defined in previous step. If we are sure that the decision we made can
actually be achieved – we are ready for the next phase.
Implementation Phase
All the previous steps we’ve made (intelligence, design, and choice) are now
implemented. Implementation can be either successful or not. Successful
implementation results with a solution to the defined problem. On the other
hand, failure brings us back to the earlier phase.
We described Simon’s model which, even today, serves as the basis of most
models of decision-making process. A process is described as a series of
events that precede final decisions. It is important to say that, at any point,
the decision maker may choose to return to the previous step for additional
validation.
Even Simon’s model was sometimes criticized as being general, that is why
we need to be aware of the importance of decision-making. This model is a
concept, a framework of how organizations and managers make decisions.
Neural networks are complex systems that mimic some features of the
functioning of the human brain. It is composed of an input layer, one or more
hidden layers, and an output layer made up of layers of artificial neurons
that are coupled. The two stages of the basic process are called
backpropagation and forward propagation.
Forward Propagation
Input Layer: Each feature in the input layer is represented by a node on
the network, which receives input data.
Weights and Connections: The weight of each neuronal connection
indicates how strong the connection is. Throughout training, these
weights are changed.
Hidden Layers: Each hidden layer neuron processes inputs by
multiplying them by weights, adding them up, and then passing them
through an activation function. By doing this, non-linearity is introduced,
enabling the network to recognize intricate patterns.
Output: The final result is produced by repeating the process until the
output layer is reached.
Backpropagation
Loss Calculation: The network’s output is evaluated against the real
goal values, and a loss function is used to compute the difference. For a
regression problem, the Mean Squared Error (MSE) is commonly used as
the cost function.
Loss Function:
Gradient Descent: Gradient descent is then used by the network to
reduce the loss. To lower the inaccuracy, weights are changed based on
the derivative of the loss with respect to each weight.
Adjusting weights: The weights are adjusted at each connection by
applying this iterative process, or backpropagation, backward across the
network.
Training: During training with different data samples, the entire process
of forward propagation, loss calculation, and backpropagation is done
iteratively, enabling the network to adapt and learn patterns from the
data.
Actvation Functions: Model non-linearity is introduced by activation
functions like the rectified linear unit (ReLU) or sigmoid. Their decision on
whether to “fire” a neuron is based on the whole weighted input.
Types of Neural Networks
There are seven types of neural networks that can be used.
Feedforward Neteworks: A feedforward neural network is a simple
artificial neural network architecture in which data moves from input to
output in a single direction. It has input, hidden, and output layers;
feedback loops are absent. Its straightforward architecture makes it
appropriate for a number of applications, such as regression and pattern
recognition.
Multilayer Perceptron (MLP): MLP is a type of feedforward neural
network with three or more layers, including an input layer, one or more
hidden layers, and an output layer. It uses nonlinear activation functions.
Convolutional Neural Network (CNN): A Convolutional Neural
Network (CNN) is a specialized artificial neural network designed for
image processing. It employs convolutional layers to automatically learn
hierarchical features from input images, enabling effective image
recognition and classification. CNNs have revolutionized computer vision
and are pivotal in tasks like object detection and image analysis.
Recurrent Neural Network (RNN): An artificial neural network type
intended for sequential data processing is called a Recurrent Neural
Network (RNN). It is appropriate for applications where contextual
dependencies are critical, such as time series prediction and natural
language processing, since it makes use of feedback loops, which enable
information to survive within the network.
Long Short-Term Memory (LSTM): LSTM is a type of RNN that is
designed to overcome the vanishing gradient problem in training RNNs. It
uses memory cells and gates to selectively read, write, and erase
information.
Support Vector Machine
Support Vector Machine (SVM) is a supervised machine learning algorithm used for both
classification and regression. Though we say regression problems as well it’s best suited for
classification. The main objective of the SVM algorithm is to find the optimal hyperplane in an N-
dimensional space that can separate the data points in different classes in the feature space. The
hyperplane tries that the margin between the closest points of different classes should be as
maximum as possible. The dimension of the hyperplane depends upon the number of features. If
the number of input features is two, then the hyperplane is just a line. If the number of input
features is three, then the hyperplane becomes a 2-D plane. It becomes difficult to imagine when
the number of features exceeds three.
Let’s consider two independent variables x1, x2, and one dependent variable which is either a blue
circle or a red circle.
From the figure above it’s very clear that there are multiple lines (our
hyperplane here is a line because we are considering only two input features
x1, x2) that segregate our data points or do a classification between red and
blue circles. So how do we choose the best line or in general the best
hyperplane that segregates our data points?
One reasonable choice as the best hyperplane is the one that represents the
largest separation or margin between the two classes.
Advantages of SVM
Euclidean Distance
This is nothing but the cartesian distance between the two points which are
in the plane/hyperplane. Euclidean distance can also be visualized as the
length of the straight line that joins the two points which are into
consideration. This metric helps us calculate the net displacement done
between the two states of an object.
Manhattan Distance
Minkowski Distance
We can say that the Euclidean, as well as the Manhattan distance, are
special cases of the Minkowski distance.
The k data points with the smallest distances to the target point are the
nearest neighbors.
This depends on the polarity base. This category can be designed as very
positive, positive, neutral, negative, or very negative. The rating is done on a
scale of 1 to 5. If the rating is 5 then it is very positive, 2 then negative, and
3 then neutral.
Emotion detection
The sentiments happy, sad, angry, upset, jolly, pleasant, and so on come
under emotion detection. It is also known as a lexicon method of sentiment
analysis.
Preprocessing
Starting with collecting the text data that needs to be analysed for
sentiment like customer reviews, social media posts, news articles, or any
other form of textual content. The collected text is pre-processed to clean
and standardize the data with various tasks:
Removing irrelevant information (e.g., HTML tags, special characters).
Tokenization: Breaking the text into individual words or tokens.
Removing stop words (common words like “and,” “the,” etc. that don’t
contribute much to sentiment).
Stemming or Lemmatization: Reducing words to their root form.
Analysis
Rule-based
Over here, the lexicon method, tokenization, and parsing come in the rule-
based. The approach is that counts the number of positive and negative
words in the given dataset. If the number of positive words is greater than
the number of negative words then the sentiment is positive else vice-versa.
Machine Learning
This approach works on the machine learning technique. Firstly, the datasets
are trained and predictive analysis is done. The next process is the
extraction of words from the text is done. This text extraction can be done
using different techniques such as Naive Bayes, Support Vector
machines, hidden Markov model , and conditional random fields like this
machine learning techniques are used.
Neural Network
In the last few years neural networks have evolved at a very rate. It involves
using artificial neural networks, which are inspired by the structure of the
human brain, to classify text into positive, negative, or neutral sentiments. it
has Recurrent neural networks, Long short-term memory , Gated recurrent
unit, etc to process sequential data like text.
Hybrid Approach
Social media posts often contain some of the most honest opinions
about your products, services, and businesses because they’re
unsolicited.
With the help of sentiment analysis software, you can wade through
all that data in minutes, to analyze individual emotions and overall
public sentiment on every social platform.
Comment 1
“Love the user interface. Setup took five minutes and we were
ready to go.”
Comment 2
“Took me 2 hours to set up, then I find out I have to update my OS.
Love it!”
Sentiment analysis would classify the second comment as negative,
even though they both use words that, without context, would be
considered positive.
Customer support
Not only that, you can keep track of your brand’s image and
reputation over time or at any given moment, so you can monitor
your progress. Whether monitoring news stories, blogs, forums, and
social media for information about your brand, you can transform
this data into usable information and statistics.
You can also trust machine learning to follow trends and anticipate
outcomes, to stay ahead and go from reactive to proactive.
Product analysis
Find out what the public is saying about a new product right after
launch, or analyze years of feedback you may have never seen. You
can search keywords for a particular product feature (interface, UX,
functionality) and use aspect-based sentiment analysis to find only
the information you need.
Use sentiment analysis for market and competitor research. Find out
who’s receiving positive mentions among your competitors, and
how your marketing efforts compare.
API Data - Data can be uploaded through Live APIs for social media. A news
API can help you glean information from all kinds of news publishers, while a
Facebook API can allow you to take all the publicly available data you need
from its platform. You can also use open source repositories like Kaggle,
or Amazon reviews.
Manual - If you have data that you already have from a CRM tool, you can
manually upload that onto the sentiment analysis API as a .csv file.
The processing of the data will depend on the kind of information it has - text,
image, video, or audio. Repustate IQ sentiment analysis steps also include
handling video content analysis with the same ease it does text analytics.
Below are the sub-tasks.
Audio transcription - The audio from the video data is transcribed through
speech to text software to ensure that any video or audio file (eg. podcast) in
the data is not overlooked.
Caption overlay - If there are any captions appearing in the video, they are
extracted by Repustate IQ and analyzed for any appearing entities, aspects,
or topics that you have identified as important.
Image overlay - Similarly, the platform recognizes and captures any images
in the video or text data through OCR (optical character recognition)
Logo recognition - Repustate IQ’s intelligent data scanner immediately
recognizes any logos that appear in the video background. This means even
videos that appear on the clothes of the presenter or say on an item like a
pen, or mug on the desk. It even picks up logos from background posters. It
does it in such a way so that not even the smallest detail goes unnoticed
when the platform conducts sentiment analysis of your brand.
Text extraction- All the text is similarly recognized and extracted in the
sentiment analysis process. This includes emojis and hashtags as well, which
are a vital part of social media sentiment analysis. Unlike other sentiment
analysis platforms, Repustate IQ ensures that emojis are never left out of data
processing because that could lead to false positives or negatives.
There are many subtasks that need to be done for this stage of the sentiment
analysis process.
Training the model - A set of dedicated, classified and labeled sentiment
analysis of dataset that will be used to train the model needs to be pre-
processed and manually labelled. It is this labelled data that will be used to
train the model by comparing the correctly classified data with the incorrectly
classified one. This will help improve the custom model that is created for a
brand.
Multilingual data - In sentiment analysis steps that include multilingual data
processing, Repustate IQ has the dataset for each language individually
annotated and trained. This is because the platform does not rely on
translations at all since all the information can get distorted, or nuances lost
due to vast differences in certain languages like Spanish and Korean. This is
the reason why Repustate gives the highest accuracy scores compared to
other platforms.
Custom Tags - In this part of the process, custom tags for aspects and
themes will be created for the data such as brand mentions, product name,
etc. Once the model has been trained, it will automatically segregate text
based on these custom created tags.
Topic Classification - The topic classifier attaches a theme to a text. For
example, this text “The dresses were awesome, and I found some really good
scarves as well.” will be tagged as the topic “clothes”.
Sentiment Analysis - Each aspect and theme is isolated in this stage by the
platform and then analysed for the sentiment. Sentiment scores are given in
the range of -1 to +1. A neutral statement may be termed as zero. This
assigning of polarity is important as ultimately, even as the platform assigns
the different scores to different aspects like convenience, speed, cleanliness,
functionality, drinks, ambience, etc, it is the aggregate score that is calculated
to know the sentiment of the audience towards the brand. So, if 3 of the 7
aspects receive a poor rating (-.65) and 4 of them receive good ones (+.5),
the sentiment analyzer will give an average score as the overall sentiment of
the brand.
Once all the steps in the sentiment analysis process have been covered, the
insights are quickly turned into actionable reports in the form of graphs and
charts. These reports can then be shared within teams as well. These visual
reports are really important because it is through them that you see granular,
aspect-based results. For example, when you get an average score for your
brand, you can filter the results in the sentiment analysis dashboard to see
which aspects got how high a score and which ones got low scores. This will
give you an idea as to what areas need your attention more than others.
Thus, at this stage of the sentiment analysis steps, you get actionable insights
that you can use to decide the right course of action for your growth plans.
Speech analytics enables the examination of customer interactions to extract meaningful insights
and customer sentiment by converting spoken words into structured data points. Leveraging
natural language processing (NLP), automatic speech recognition (ASR), machine learning, and
artificial intelligence (AI), it extracts customer preferences, behavior, and emotions from customer
conversations.