DAIOT UNIT 5 (1) Own
DAIOT UNIT 5 (1) Own
DAIOT UNIT 5 (1) Own
Machine Learning tutorial provides basic and advanced concepts of machine learning.
Our machine learning tutorial is designed for students and working professionals.
This machine learning tutorial gives you an introduction to machine learning along
with the wide range of machine learning techniques such
as Supervised, Unsupervised, and Reinforcement learning. You will learn about
regression and classification models, clustering methods, hidden Markov models, and
various sequential models.
With the help of sample historical data, which is known as training data, machine
learning algorithms build a mathematical model that helps in making predictions or
decisions without being explicitly programmed. Machine learning brings computer
science and statistics together for creating predictive models. Machine learning
constructs or uses the algorithms that learn from historical data. The more we will
provide the information, the higher will be the performance.
A machine has the ability to learn if it can improve its performance by gaining
more data.
We can train machine learning algorithms by providing them the huge amount of data
and let them explore the data, construct the models, and predict the required output
automatically. The performance of the machine learning algorithm depends on the
amount of data, and it can be determined by the cost function. With the help of
machine learning, we can save both time and money.
The importance of machine learning can be easily understood by its uses cases,
Currently, machine learning is used in self-driving cars, cyber fraud detection, face
recognition, and friend suggestion by Facebook, etc. Various top companies such
as Netflix and Amazon have build machine learning models that are using a vast
amount of data to analyze the user interest and recommend product accordingly.
Following are some key points which show the importance of Machine Learning:
1. Supervised learning
2. Unsupervised learning
3. Reinforcement learning
1) Supervised Learning
Supervised learning is a type of machine learning method in which we provide sample
labeled data to the machine learning system in order to train it, and on that basis, it
predicts the output.
The system creates a model using labeled data to understand the datasets and learn
about each data, once the training and processing are done then we test the model
by providing a sample data to check whether it is predicting the exact output or not.
The goal of supervised learning is to map input data with the output data. The
supervised learning is based on supervision, and it is the same as when a student learns
things in the supervision of the teacher. The example of supervised learning is spam
filtering.
o Classification
o Regression
2) Unsupervised Learning
Unsupervised learning is a learning method in which a machine learns without any
supervision.
The training is provided to the machine with the set of data that has not been labeled,
classified, or categorized, and the algorithm needs to act on that data without any
supervision. The goal of unsupervised learning is to restructure the input data into new
features or a group of objects with similar patterns.
o Clustering
o Association
3) Reinforcement Learning
Reinforcement learning is a feedback-based learning method, in which a learning
agent gets a reward for each right action and gets a penalty for each wrong action.
The agent learns automatically with these feedbacks and improves its performance. In
reinforcement learning, the agent interacts with the environment and explores it. The
goal of an agent is to get the most reward points, and hence, it improves its
performance.
The robotic dog, which automatically learns the movement of his arms, is an example
of Reinforcement learning.
One of the best ways to dramatically improve the predictive ability of your ML models
is not in the algorithms themselves, but in how the data that they are grown from is
presented to them. The transformations of data, the addition of constructed new
fields, and the removal of distracting fields is all done with the knowledge of how the
representation model operates. This process is called feature engineering. Data
fields are commonly referred to as features in ML. We will adopt that terminology for
the rest of the chapter.
The goal of feature engineering is to make it as easy as possible for your ML model
to have good performance. Different representations have different requirements for
what works well, so you will find yourself creating different versions of the same raw
dataset geared specifically to the ML representation. Get to know each ML
representation you are using to make sure you are giving it the best possible chance
to perform.
Feature engineering is an art and it is hard. But, it can add a lot of value and greatly
increase your probability of success. We will introduce a few key concepts, but there
is much, much more to learn.
❤...: Analytical method validation is a process used to ensure that the analytical method used for
a particular test is suitable for its intended use.
❤...: method validation results are used to ensure the quality, reliability and consistency of
analytical results; This is an integral part of any good analytical practice.
❤...: Validation should be done according to the validation protocol. Protocol should have
procedures and acceptance criteria for all roles
❤...: The results should be recorded in the validation report. Justification should be provided
when non-pharmacopoeial methods are used.
❤...
- Accuracy
- Firmness
- Simplicity
- Range
- Uniqueness
- Detection limits
- Size limit
❤...:1. Accuracy
Accuracy must be established within a certain range of the analytical procedure.
❤...: 2. Accuracy
It is the degree of agreement between individual results.
❤...: 3. Firmness
This should be considered at the development stage and show the reliability of the analysis when
deliberate variations are made in the method parameters.
❤...: 4. Simplicity.
It refers to the ability to produce results that are directly proportional to the concentration of the
analyte in the samples.
❤...: 5. Scope
It is an expression of the lowest and highest levels of analyte presented as determinable for the
product. The specified range is usually derived from linearity studies.
❤...: 6. Selection
It is the ability to unambiguously measure the desired analyte in the presence of components such
as excipients and impurities.
❤...: 7. Limitation of detection (limitation of detection)
It is the smallest quantity of analyte that can be detected by quantitative method and is not
necessarily determind.
❤...: 8. Quantity Limitation (Quantity Limitation)
It is the lowest concentration of an analyte in a sample that can be determined with acceptable
precision and accuracy.
• A model with a higher bias would not match the data set
closely.
• A low bias model will closely match the training data set.
Bias-Variance Trade-Off
While building the machine learning model, it is really important to take care of bias
and variance in order to avoid overfitting and underfitting in the model. If the model
is very simple with fewer parameters, it may have low variance and high bias. Whereas,
if the model has a large number of parameters, it will have high variance and low bias.
So, it is required to make a balance between bias and variance errors, and this balance
between the bias error and variance error is known as the Bias-Variance trade-off.
For an accurate prediction of the model, algorithms need a low variance and low bias.
But this is not possible because bias and variance are related to each other:
Hence, the Bias-Variance trade-off is about finding the sweet spot to make a
balance between bias and variance errors.
o High training error and the test error is almost similar to training error
Deep learning can do wonders for complex data, with thousands to millions of features
and a large history of labeled examples to use as training sets. Rapid advances in image
recognition have as much to do with the vast trove of recognized images that Google
and others have amassed over the years as it does with advances in the deep learning
algorithms used.
For loT data, this limits the usefulness of deep learning techniques. Most loT data is
relatively new without a long history of labeled examples. Most IoT devices only have
a few sensors, so the feature set is not that complex. In these situations, many of the
previously discussed ML techniques can perform a predictive job as well, if not better,
than deep learning techniques. Deep learning is computationally expensive, both in
terms of time and computational power (i.e. high cost). However, the loT data flow
consists of a large number of features and hundred With thousands to millions of
labeled training data available, deep learning methods can provide a significant boost
in predictive power. It is clear in Autonomous vehicle development. It can also work
wonders where pictures are taken The device is part of the activity (static or video).
Deep learning packages interact better with Python versus R. They're also relatively
new, so expect documentation and tutorials to be limited. This makes them a little
trickier to work with than the previous examples in this chapter. You need more time
and expertise to develop deep learning models than well-run ML models like Random
Forest.
Use whichever method makes the most sense for the individual use case and available
training data. Consider the impact requirements and see if using a deep learning
modeling technique provides enough accuracy to warrant the cost.