Naveen Pandey has more than 2 years of experience in data science and machine learning. He is an experienced Machine Learning Engineer with a strong background in data analysis, natural language processing, and machine learning. Holding a Bachelor of Science in Information Technology from Sikkim Manipal University, he excels in leveraging cutting-edge technologies such as Large Language Models (LLMs), TensorFlow, PyTorch, and Hugging Face to develop innovative solutions.
The ROC curve is a graph between False positive rate on the x axis and True positive rate on the y axis. True positive rate is the ratio of True positives to the total number of positive samples. False positive rate is the ratio of False positives to the total number of negative samples. The…
True Positive rate (TRP) is the ratio of True Positives to True Positives and False Negatives. It is the probability that an actual positive will test as positive. TPR = TP / TP + FN The False Positive Rate (FPR) is the ratio of the False Positives to all the positives (True positives and false…
1. What is the significance of sampling? Name some techniques for sampling? For analyzing the data, we cannot proceed with the whole volume at once for large datasets. We need to take some samples from the data which can represent the whole population. While making a sample out of complete data, we should take the…
1. Perceptron The perceptron is the most basic of all neural networks, being a fundamental building block of more complex neural network. If simple connects an input cell and an output cell. 2. Feed-Forward Network The feed-forward network is a collection of perceptions’. In which there are three fundamental types of layers – input layers,…
1. Bagging and Boosting Bagging and Boosting are two different ways used in combining base estimators for ensemble learning (Like random forest combining decision trees). Bagging means aggregating the predictions of several weak learners. We can think of it combining weak learners is used in parallel. The average of the predictions of several weak learners…
Probability distribution is the function that shows the probabilities of the outcome of an event or experiment. Consider a feature (i.e., column) in a dataframe. This feature is a variable and its probability distribution function shows the likelihood of the values it can take. Probability distribution function are quite useful in predictive analytics or machine…
Converge Algorithm that converges will eventually reach an optimal answer, even if very slowly. An algorithm that doesn’t converge may never reach an optimal answer. Learning Rate Rate at which optimizers change weights and biases. High learning rate generally trains faster but risks not converging whereas a lower rate trains slower. Numerical instability Issues with…
Neuron Node is a NN, typically taking in multiple input values and generating one output value by applying an activation function (nonlinear transformation) to weighted sum of input values. Weights Edges is a NN, the goal of training is to determine the optimal weight for each feature; if weight = 0, corresponding feature does not…
In this article I will take you through the task of Analyzing the Russia-Ukraine war Dataset using Python. The dataset that I am using for the task of analysis the Ukraine and Russia War is downloaded from Kaggle. You can download russia-ukraine equipment dataset from here and russia-ukraine personnel losses dataset from here. Now let’s import…
Central Limit Theorem We first need to introduce the normal (gaussian) distribution for central limit theorem to make sense. Normal distribution is a probability distribution that look like a bell. X-axis represents the values and y-axis represents the probability of observing these values. The sigma values represent standard deviation normal distribution is used to represent…
Definition Feature scaling is the process of normalizing the range of feature in a dataset. Real-world datasets often contain features that are varying in degrees of magnitude, range and units. Therefore, in order for machine learning models to interpret these features on the same scale, we need to perform scaling. Feature scaling makes the model…
Convolutional Neural Networks CNN’s popularly known as ConvNets majority consists of several layers and are specifically used for image processing and detection of objects. It was developed in 1998 by Yann LeCun. CNNs have wide usage in identifying the image of the satellites, medical image processing, series forecasting, and anomaly detection. CNNs process the data…