{"id":11290,"date":"2018-10-01T09:49:12","date_gmt":"2018-10-01T04:19:12","guid":{"rendered":"http:\/\/ucanalytics.com\/blogs\/?p=11290"},"modified":"2019-04-21T18:30:41","modified_gmt":"2019-04-21T13:00:41","slug":"math-of-deep-learning-neural-networks-simplified-part-2","status":"publish","type":"post","link":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/","title":{"rendered":"Math of Deep Learning Neural Networks &#8211; Simplified (Part 2)"},"content":{"rendered":"<hr \/>\n<div id=\"attachment_11291\" style=\"width: 806px\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg\"><img aria-describedby=\"caption-attachment-11291\" data-attachment-id=\"11291\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/deep-learning-neural-networks-and-plumbing-job\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&amp;ssl=1\" data-orig-size=\"796,597\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Deep Learning Neural Networks and Plumbing Job\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=300%2C225&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=640%2C480&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-11291 size-full\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?resize=640%2C480\" alt=\"\" width=\"640\" height=\"480\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?w=796&amp;ssl=1 796w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?resize=250%2C188&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?resize=300%2C225&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?resize=768%2C576&amp;ssl=1 768w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a><p id=\"caption-attachment-11291\" class=\"wp-caption-text\">The Math of Deep Learning Neural Networks &#8211; by Roopam<\/p><\/div>\n<p>Welcome back to this series of articles on <a href=\"http:\/\/ucanalytics.com\/blogs\/category\/deep-learning-neural-networks\/\" rel=\"noopener\">deep learning and neural networks<\/a>. In the last part, you learned how training a\u00a0<a href=\"http:\/\/ucanalytics.com\/blogs\/deep-learning-and-neural-networks-simplified-part-1\/\" rel=\"noopener\">deep learning network is similar to a plumbing job<\/a><strong>. <\/strong>This time you will learn the math of deep learning. We will continue to use the plumbing analogy to simplify the seemingly complicated math. I believe you will find this highly intuitive. Moreover, understanding this will provide you with a good idea about the inner workings of deep learning networks and artificial intelligence (AI) to build an AI of your own. We will use the math of deep learning to make an image recognition AI in the next part. But before that let&#8217;s create the links between&#8230;<\/p>\n<h2><span style=\"color: #3366ff;\">The Math of Deep Learning and Plumbing<\/span><\/h2>\n<p>Last time we noticed that neural networks are like the networks of water pipes. The goal of neural networks is to identify the right settings for the knobs (6 in this schematic) to get the\u00a0right output given the input.<\/p>\n<p><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg\"><img data-attachment-id=\"11228\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/deep-learning-and-neural-networks-simplified-part-1\/deep-learning-neural-networks-water-pipeline-1\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?fit=1896%2C789&amp;ssl=1\" data-orig-size=\"1896,789\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Deep Learning Neural Networks &#8211; Water Pipeline 1\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?fit=300%2C125&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?fit=640%2C266&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-11228 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?resize=640%2C266\" alt=\"\" width=\"640\" height=\"266\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?w=1896&amp;ssl=1 1896w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?resize=250%2C104&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?resize=300%2C125&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?resize=768%2C320&amp;ssl=1 768w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?resize=1024%2C426&amp;ssl=1 1024w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Water-Pipeline-1.jpg?w=1280 1280w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a><\/p>\n<p>Shown below is a familiar schematic of neural networks almost identical to the water pipelines above. The only exception is the additional bias terms (b<sub>1<\/sub>,b<sub>2<\/sub>, and b<sub>3<\/sub>) added to the nodes.<\/p>\n<p><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg\"><img data-attachment-id=\"11320\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/maths-of-deep-learning-neural-networks-image-3\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=1470%2C840&amp;ssl=1\" data-orig-size=\"1470,840\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Maths of Deep Learning Neural Networks (Image 3)\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=300%2C171&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=640%2C366&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-11320 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=640%2C366\" alt=\"\" width=\"640\" height=\"366\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?w=1470&amp;ssl=1 1470w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=250%2C143&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=300%2C171&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=768%2C439&amp;ssl=1 768w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=1024%2C585&amp;ssl=1 1024w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?w=1280 1280w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a>In this post, we will solve this network to understand the math of deep learning. Note that a deep learning model has multiple hidden layers, unlike this simple neural network. However, this simple neural network can easily be generalized to the deep learning models. The math of deep learning does not change a lot with additional complexity and hidden layers. Here, our objective is to identify the values of the parameters {W (W<sub>1<\/sub>,&#8230;, W<sub>6<\/sub>) and b (b<sub>1<\/sub>,b<sub>2<\/sub>, and b<sub>3<\/sub>)}. We will soon use the backpropagation algorithm along with <strong><a href=\"http:\/\/ucanalytics.com\/blogs\/category\/gradient-descent\/\" rel=\"noopener\">gradient descent optimization<\/a><\/strong> to solve this network and identify the optimal\u00a0values of these weights.<\/p>\n<p><img data-attachment-id=\"11245\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/deep-learning-and-neural-networks-simplified-part-1\/backpropagation\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?fit=414%2C308&amp;ssl=1\" data-orig-size=\"414,308\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Backpropagation\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?fit=300%2C223&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?fit=414%2C308&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-11245 alignright\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?resize=275%2C205\" alt=\"\" width=\"275\" height=\"205\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?w=414&amp;ssl=1 414w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?resize=250%2C186&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Backpropagation.jpg?resize=300%2C223&amp;ssl=1 300w\" sizes=\"(max-width: 275px) 100vw, 275px\" data-recalc-dims=\"1\" \/><\/p>\n<h2><span style=\"color: #3366ff;\">Backpropagation and Gradient Descent<\/span><\/h2>\n<p>In the previous post, we discussed that the backpropagation algorithm works similar to me shouting back at my plumber while he was working in the duct. Remember, I was telling the plumber about the difference in actual water pressure from the expected. The plumber of neural networks, unlike my building&#8217;s plumber,\u00a0learns from this information to optimize the positions of the knobs. The method that the neural networks plumber uses to iteratively correct the weights or settings of the knobs is called gradient descent.<\/p>\n<p>We have discussed the <a href=\"http:\/\/ucanalytics.com\/blogs\/gradient-descent-logistic-regression-simplified-step-step-visual-guide\/\" rel=\"noopener\"><strong>gradient descent algorithm<\/strong><\/a> in an earlier post to solve a logistic regression model. I recommend that you read that article to get a good grasp of the things we will discuss in this post. Essentially, the idea is to iteratively correct the value of the weights (W<sub>i<\/sub>) to produce the least difference between the actual and the expected values of the output. This difference is measured mathematically by the loss function i.e <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cell+f+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ell f \" class=\"latex\" \/>. The weights (W<sub>i<\/sub> and b<sub>i<\/sub>) are then iteratively improved using the gradient of the loss function wrt weights using this expression:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=W_%7Bi%7D+%5Cleftarrow%C2%A0+W_%7Bi%7D+%2B+%5Calpha+%5Ccdot+%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7Bi%7D%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"W_{i} &#92;leftarrow\u00a0 W_{i} + &#92;alpha &#92;cdot &#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{i}} \" class=\"latex\" \/><\/pre>\n<p>Here, \u03b1 is called the learning rate &#8211; it&#8217;s a hyperparameter and stays constant. Hence, the overall problem boils down to the identification of partial derivatives of the loss function with respect to the weights i.e. <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7Bi%7D%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{i}} \" class=\"latex\" \/>. For our problem, we just need to solve the partial derivatives for W<sub>5<\/sub> and W<sub>1<\/sub>. The partial derivatives for other weights can then be easily derived using the same method used for W<sub>5<\/sub> and W<sub>1<\/sub>.<\/p>\n<p>Before we solve these partial derivatives, let&#8217;s do some more plumbing jobs and look at a tap to develop intuitions about the results we will get from the gradient descent optimization.<\/p>\n<h2><span style=\"color: #3366ff;\">Intuitive Math of Deep Learning for W<\/span><sub><span style=\"color: #3366ff;\">5<\/span>\u00a0<\/sub><span style=\"color: #3366ff;\">&amp; A Tap<\/span><\/h2>\n<p><img data-attachment-id=\"11311\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/maths-of-deep-learning-neural-network-tap\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?fit=426%2C704&amp;ssl=1\" data-orig-size=\"426,704\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Maths of Deep Learning Neural Network Tap\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?fit=182%2C300&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?fit=426%2C704&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-11311 alignright\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?resize=280%2C462\" alt=\"\" width=\"280\" height=\"462\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?w=426&amp;ssl=1 426w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?resize=151%2C250&amp;ssl=1 151w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Network-Tap.jpg?resize=182%2C300&amp;ssl=1 182w\" sizes=\"(max-width: 280px) 100vw, 280px\" data-recalc-dims=\"1\" \/><\/p>\n<p>We will use this simple tap to identify an optimal setting for its knob. In\u00a0this process, we will develop intuitions about gradient descent and the math of deep learning. Here, the input is the water coming from the pipe on the left of the image. Moreover, the output is the water coming out of the tap. You use the knob, on the\u00a0top of the\u00a0tap, to regulate the quantity of the output water given the input.\u00a0Remember, you want to turn the knob in such a way that you get the desired output (i.e the quantity of water) to wash your hands. Keep in mind, the position of the knob is similar to the weight of a neural networks&#8217; parameters. Moreover, the input\/output water is similar to the input\/output variables.<\/p>\n<p>Essentially, in math terms, you are trying to identify how the position of the\u00a0knob\u00a0influences the output water. The mathematical equation for the same is:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%28Output%5C+water%5C+from%5C+the%5C+tap%29%7D%7B%5Cpartial+%28Position%2FSetting%5C+of%5C+the%5C+knob%29%7D%3D%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7Bi%7D%7D&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial (Output&#92; water&#92; from&#92; the&#92; tap)}{&#92;partial (Position\/Setting&#92; of&#92; the&#92; knob)}=&#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{i}}\" class=\"latex\" \/><\/pre>\n<p>If you understand the influence of the knob on the output flow of water you can easily turn it to get the desired output. Now, let&#8217;s develop an intuition about how much to twist the knob. When you use a tap you twist the knob until you get the right flow or the output. When the difference between the desired output and the actual output is large then you need a lot of twisting. On the other hand, when the difference is less then you turn the knob gently.<\/p>\n<p>Moreover, the other factor on which your decision depends on is the input from the left pipe. If there is no water flowing from the left pipe then no matter how much you twist the\u00a0knob it won&#8217;t help. Essentially, your action depends on these two factors.<\/p>\n<pre><span style=\"font-size: 14pt;\"><strong>Your decision to turn the knob depends on\n<\/strong><strong><span style=\"color: #0000ff;\"><span style=\"color: #333333;\">Factor 1:<\/span> Difference between the actual output and the desired Output <span style=\"color: #ff0000;\">and<\/span>\n<\/span><span style=\"color: #0000ff;\"><span style=\"color: #333333;\">Factor 2:<\/span> Input from the grey pipe on the left\n<\/span><\/strong><\/span><\/pre>\n<p>Soon you will get the same result by doing a seemingly complicated math for the gradient descent to solve the neural network.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7B5%7D%7D%3D%28Output%5C+Difference%29+%5Ccdot+%28Input%29+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{5}}=(Output&#92; Difference) &#92;cdot (Input) \" class=\"latex\" \/><\/pre>\n<p>For our network, the output difference is <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%28%5Cltimes_%7B3%7D-y%29&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"(&#92;ltimes_{3}-y)\" class=\"latex\" \/> and input is <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ltimes_{1} \" class=\"latex\" \/>. Hence,<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7B5%7D%7D%3D%28%5Cltimes_%7B3%7D-y%29+%5Ccdot+%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{5}}=(&#92;ltimes_{3}-y) &#92;cdot &#92;ltimes_{1} \" class=\"latex\" \/><\/pre>\n<h4><span style=\"color: #3366ff;\">Disclaimer<\/span><\/h4>\n<p>Please note,\u00a0to make the\u00a0concepts easy for you to understand, I had taken a few liberties while defining the factors in the previous section. I will make these factors much more theoretically grounded at the end of this article when I will discuss the chain rule to solve derivatives. For now, I will continue to take more liberties in the next section when I discuss the other weight modification for other parameters of neural networks.<\/p>\n<h2><span style=\"color: #3366ff;\">Add More Knobs to Solve W<sub>1<\/sub> &#8211;\u00a0Intuitive Math of Deep Learning<\/span><\/h2>\n<p>Neural networks, as discussed earlier, have several parameters (Ws and bs). To develop an intuition about the math to estimate the other parameters further away from the output (i.e. W<sub>1<\/sub>), let&#8217;s add another knob to the tap.\u00a0<a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg\"><img data-attachment-id=\"11389\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/maths-of-deep-learning-neural-networks-two-knobs\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?fit=978%2C635&amp;ssl=1\" data-orig-size=\"978,635\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Maths of Deep Learning Neural Networks &#8211; two knobs\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?fit=300%2C195&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?fit=640%2C416&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"wp-image-11389 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?resize=558%2C362\" alt=\"\" width=\"558\" height=\"362\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?w=978&amp;ssl=1 978w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?resize=250%2C162&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?resize=300%2C195&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-two-knobs.jpg?resize=768%2C499&amp;ssl=1 768w\" sizes=\"(max-width: 558px) 100vw, 558px\" data-recalc-dims=\"1\" \/><\/a><br \/>\nHere, we have added a red regulator knob to the tap we saw in the earlier section. Now, the output from the tap is governed by both these knobs. Referring to the neural network&#8217;s image shown earlier, the red knob is similar to the parameters (W<sub>1<\/sub>, W<sub>2,<\/sub> W<sub>3<\/sub>, W<sub>4, <\/sub>b<sub>1, <\/sub>and b<sub>2<\/sub>) added to the hidden layers. The knob on top of the brass tap is like the parameters to the output layer (i.e. W<sub>5<\/sub>, W<sub>6<\/sub>, and b<sub>3<\/sub>).<\/p>\n<p>Now, you are also using the red knob, in addition to the knob on the tap, to get the desired output from the tap. Your effort of the red knob will depend on these factors.<\/p>\n<pre><span style=\"font-size: 14pt;\"><strong>Your decision to turn the red knob depends on\n<\/strong><strong><span style=\"color: #0000ff;\"><span style=\"color: #333333;\">Factor 1:<\/span> Difference between the actual and the desired final output <span style=\"color: #ff0000;\">and<\/span>\n<\/span>Factor 2<span style=\"color: #0000ff;\">: Position \/ setting of the knob on the brass tap <span style=\"color: #ff0000;\">and\n<span style=\"color: #333333;\">Factor 3:<\/span> <\/span>Change in input to the brass tap caused by the red knob<span style=\"color: #ff0000;\"> and\n<span style=\"color: #333333;\">Factor 4:<\/span> <span style=\"color: #0000ff;\">Input from the pipe on the left into the red knob\n<\/span><\/span><\/span><\/strong><\/span><\/pre>\n<p>Here, as already discussed earlier, factor 1 is <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%28%5Cltimes_%7B3%7D-y%29&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"(&#92;ltimes_{3}-y)\" class=\"latex\" \/>. W<sub>5<\/sub> is the setting\/weight for the knob of the brass tap. Factor 3 is <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%28%5Cltimes_%7B1%7D-1%29%5Ccdot%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"(&#92;ltimes_{1}-1)&#92;cdot&#92;ltimes_{1} \" class=\"latex\" \/>. Finally, the last factor is the input or X<sub>1<\/sub>. This completes our equation as:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f%7D%7B%5Cpartial+W_%7B1%7D%7D%3D%28%5Cltimes_%7B3%7D-y%29%5Ccdot+W_%7B5%7D+%5Ccdot+%28%5Cltimes_%7B1%7D-1%29%5Ccdot%5Cltimes_%7B1%7D%5Ccdot+X_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f}{&#92;partial W_{1}}=(&#92;ltimes_{3}-y)&#92;cdot W_{5} &#92;cdot (&#92;ltimes_{1}-1)&#92;cdot&#92;ltimes_{1}&#92;cdot X_{1} \" class=\"latex\" \/><\/pre>\n<p>Now, before we do the math to get these results, we just need to discuss the components of our neural network in mathematical terms. We already know how it relates to the water pipelines discussed earlier.<\/p>\n<p><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg\"><img data-attachment-id=\"11320\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/maths-of-deep-learning-neural-networks-image-3\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=1470%2C840&amp;ssl=1\" data-orig-size=\"1470,840\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Maths of Deep Learning Neural Networks (Image 3)\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=300%2C171&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?fit=640%2C366&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-11320 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=640%2C366\" alt=\"\" width=\"640\" height=\"366\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?w=1470&amp;ssl=1 1470w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=250%2C143&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=300%2C171&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=768%2C439&amp;ssl=1 768w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?resize=1024%2C585&amp;ssl=1 1024w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Maths-of-Deep-Learning-Neural-Networks-Image-3.jpg?w=1280 1280w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a><\/p>\n<p>Let&#8217;s start with the nodes or the orange circles in the network diagram.<\/p>\n<h2><span style=\"color: #3366ff;\">Nodes of\u00a0<\/span><span style=\"color: #3366ff;\">Neural Networks<\/span><\/h2>\n<p>Here, these two networks are equivalent except the additional b<sub>1<\/sub> or bias for the neural networks.<\/p>\n<p><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg\"><img data-attachment-id=\"11523\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/?attachment_id=11523\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?fit=1268%2C428&amp;ssl=1\" data-orig-size=\"1268,428\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Single Input Node &#8211; Maths of Deep Learning Neural Network\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?fit=300%2C101&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?fit=640%2C216&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-11523 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?resize=640%2C216\" alt=\"\" width=\"640\" height=\"216\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?w=1268&amp;ssl=1 1268w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?resize=250%2C84&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?resize=300%2C101&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?resize=768%2C259&amp;ssl=1 768w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Single-Input-Node-Maths-of-Deep-Learning-Neural-Network.jpg?resize=1024%2C346&amp;ssl=1 1024w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a><\/p>\n<p>The node for the neural network has two components i.e. sum and non-linear. The sum component (Z<sub>1<\/sub>) is just a linear combination of the input and the weights.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=Z_%7B1%7D+%3D+W_%7B1%7D%5Ccdot+X_%7B1%7D+%2B%C2%A0W_%7B2%7D%5Ccdot+X_%7B2%7D+%2Bb_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"Z_{1} = W_{1}&#92;cdot X_{1} +\u00a0W_{2}&#92;cdot X_{2} +b_{1} \" class=\"latex\" \/><\/pre>\n<p>The next term, i.e. non-linear, is the non-linear sigmoid activation function (<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ltimes_{1} \" class=\"latex\" \/>). As discussed earlier, it is like a regulator of a fan that keeps the value of <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ltimes_{1} \" class=\"latex\" \/>\u00a0 between 0 and 1 or on\/off.<\/p>\n<p><a href=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg\"><img data-attachment-id=\"11224\" data-permalink=\"https:\/\/ucanalytics.com\/blogs\/deep-learning-and-neural-networks-simplified-part-1\/neural-networks-activation-functions-sigmoid-function\/\" data-orig-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?fit=1453%2C757&amp;ssl=1\" data-orig-size=\"1453,757\" data-comments-opened=\"1\" data-image-meta=\"{&quot;aperture&quot;:&quot;0&quot;,&quot;credit&quot;:&quot;&quot;,&quot;camera&quot;:&quot;&quot;,&quot;caption&quot;:&quot;&quot;,&quot;created_timestamp&quot;:&quot;0&quot;,&quot;copyright&quot;:&quot;&quot;,&quot;focal_length&quot;:&quot;0&quot;,&quot;iso&quot;:&quot;0&quot;,&quot;shutter_speed&quot;:&quot;0&quot;,&quot;title&quot;:&quot;&quot;,&quot;orientation&quot;:&quot;0&quot;}\" data-image-title=\"Neural Networks Activation Functions &#8211; Sigmoid Function\" data-image-description=\"\" data-image-caption=\"\" data-medium-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?fit=300%2C156&amp;ssl=1\" data-large-file=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?fit=640%2C333&amp;ssl=1\" decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-11224 aligncenter\" src=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?resize=640%2C333\" alt=\"\" width=\"640\" height=\"333\" srcset=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?w=1453&amp;ssl=1 1453w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?resize=250%2C130&amp;ssl=1 250w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?resize=300%2C156&amp;ssl=1 300w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?resize=768%2C400&amp;ssl=1 768w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?resize=1024%2C533&amp;ssl=1 1024w, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Neural-Networks-Activation-Functions-Sigmoid-Function.jpg?w=1280 1280w\" sizes=\"(max-width: 640px) 100vw, 640px\" data-recalc-dims=\"1\" \/><\/a><\/p>\n<p>The mathematical expression for this sigmoid activation function (<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ltimes_{1} \" class=\"latex\" \/>) is:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B1%7D+%3D+%5Cfrac%7Be%5E%7BZ_%7B1%7D%7D%7D%7B1%2Be%5E%7BZ_%7B1%7D%7D%7D+&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;ltimes_{1} = &#92;frac{e^{Z_{1}}}{1+e^{Z_{1}}} \" class=\"latex\" \/><\/pre>\n<p>The nodes in both the hidden and output layer behave the same as described above. Now, the last thing is to define the loss function (<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cell+f&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ell f\" class=\"latex\" \/>) which is to measure the difference between the\u00a0expected and actual output. We will define the loss function for most common business problems.<\/p>\n<h2><span style=\"color: #3366ff;\">Classification Problem &#8211;\u00a0Math of Deep Learning<\/span><\/h2>\n<p>In practice, most business problems are about classification. They have binary or categorical outputs\/answers such as:<\/p>\n<ul style=\"list-style-type: square;\">\n<li>Is the last credit card transaction fraudulent or not?<\/li>\n<li>Will the borrower return the money or not?<\/li>\n<li>Was the last email in your mailbox a spam or ham?<\/li>\n<li>Is that a picture of a dog or cat? (this is not a business problem but a famous problem for deep learning)<\/li>\n<li>Is there an object in front of an autonomous car to generate a signal to push the break?<\/li>\n<li>Will the person surfing the web respond to the ad of a luxury condo?<\/li>\n<\/ul>\n<p>Hence, we will design the loss function of our neural network for similar binary outputs. This binary loss function, aka binary cross entropy, can easily be\u00a0extended for multiclass problems with minor modifications.<\/p>\n<h2><span style=\"color: #3366ff;\">Loss Function and Cross Entropy<\/span><\/h2>\n<p>The loss function for binary output problems is:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cell+f+%3D-%281-y%29%5Ccdot+ln%281-%5Cltimes_%7B3%7D%29-y%5Ccdot+ln%28%5Cltimes_%7B3%7D%29&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;ell f =-(1-y)&#92;cdot ln(1-&#92;ltimes_{3})-y&#92;cdot ln(&#92;ltimes_{3})\" class=\"latex\" \/><\/pre>\n<p>This expression is also referred to as binary cross entropy. We can easily extend this binary cross-entropy to multi-class entropy if the output has many classes such as images of dog, cat, bus, car etc. We will learn about multiclass cross entropy and softmax function in the next part of this series. Now that we have identified all the components of the neural network, we are ready to solve it using the chain rule of differential equations.<\/p>\n<h2><span style=\"color: #3366ff;\">Chain Rule for W<sub>5\u00a0<\/sub><\/span><span style=\"color: #3366ff;\">&#8211; Math of Deep Learning<\/span><\/h2>\n<p>We discussed the outcome for change observed in the loss function(<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cell+f+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ell f \" class=\"latex\" \/>) wrt to change in W<sub>5 <\/sub>earlier using a single knob analogy. We know the answer to\u00a0<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+W_%7B5%7D%7D&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial W_{5}}\" class=\"latex\" \/> is equal to\u00a0\u00a0<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%28%5Cltimes_%7B3%7D-y%29%5Ccdot+%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"(&#92;ltimes_{3}-y)&#92;cdot &#92;ltimes_{1} \" class=\"latex\" \/>. Now, let&#8217;s derive the same thing using the chain rule of derivatives. Essentially, this is similar to the change in water pressure observed at the output by turning the knob on the top of the tap. The chain rule states this:<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+W_%7B5%7D%7D%3D+%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+%5Cltimes_%7B3%7D%7D+%5Ccdot+%5Cfrac%7B%5Cpartial+%5Cltimes_%7B3%7D%7D%7B%5Cpartial+Z_%7B3%7D%7D%5Ccdot+%5Cfrac%7B%5Cpartial+Z_%7B3%7D%7D%7B%5Cpartial+W_%7B5%7D%7D&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial W_{5}}= &#92;frac{&#92;partial &#92;ell f }{&#92;partial &#92;ltimes_{3}} &#92;cdot &#92;frac{&#92;partial &#92;ltimes_{3}}{&#92;partial Z_{3}}&#92;cdot &#92;frac{&#92;partial Z_{3}}{&#92;partial W_{5}}\" class=\"latex\" \/><\/pre>\n<p>The above equation for chain rule is fairly simple since equation on the right-hand side will become the one on the left-hand side by simple division. More importantly, these equations suggest that the change in the output essentially the change observed at different components of the pipeline because of turning\u00a0the knob.<\/p>\n<p>Moreover, we already discussed the loss function which is the binary cross entropy i.e.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cell+f+%3D-%281-y%29%5Ccdot+ln%281-%5Cltimes_%7B3%7D%29-y%5Ccdot+ln%28%5Cltimes_%7B3%7D%29&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;ell f =-(1-y)&#92;cdot ln(1-&#92;ltimes_{3})-y&#92;cdot ln(&#92;ltimes_{3})\" class=\"latex\" \/><\/pre>\n<p>The first component of the chain rule is\u00a0<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+%5Cltimes_%7B3%7D%7D&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial &#92;ltimes_{3}}\" class=\"latex\" \/> which is<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+%5Cltimes_%7B3%7D%7D%3D%5Cfrac%7B%281-y%29%7D%7B1-%5Cltimes_%7B3%7D%7D-%5Cfrac%7By%7D%7B%5Cltimes_%7B3%7D%7D%3D%5Cfrac%7B%28%5Cltimes_%7B3%7D-y%29%7D%7B%5Cltimes_%7B3%7D%5Ccdot%281-%5Cltimes_%7B3%7D%29%7D&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial &#92;ltimes_{3}}=&#92;frac{(1-y)}{1-&#92;ltimes_{3}}-&#92;frac{y}{&#92;ltimes_{3}}=&#92;frac{(&#92;ltimes_{3}-y)}{&#92;ltimes_{3}&#92;cdot(1-&#92;ltimes_{3})}\" class=\"latex\" \/><\/pre>\n<p>This was fairly easy to compute if you only know that derivative of a natural log function is<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+ln%28x%29%7D%7B%5Cpartial+x%7D%3D%5Cfrac%7B1%7D%7Bx%7D&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial ln(x)}{&#92;partial x}=&#92;frac{1}{x}\" class=\"latex\" \/><\/pre>\n<p>This second component of the step function is\u00a0<img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cltimes_%7B3%7D%7D%7B%5Cpartial+Z_%7B3%7D%7D&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ltimes_{3}}{&#92;partial Z_{3}}\" class=\"latex\" \/>. This derivative of the sigmoid function ( <img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cltimes_%7B3%7D&#038;bg=ffffff&#038;fg=000&#038;s=0&#038;c=20201002\" alt=\"&#92;ltimes_{3}\" class=\"latex\" \/>) is slightly more complicated.\u00a0 You could find here a detailed solution to\u00a0<strong><a href=\"http:\/\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Derivative-of-Sigmoid.jpg\" rel=\"noopener\">the derivative of the sigmoid function<\/a>. <\/strong>This implies,<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cltimes_%7B3%7D%7D%7B%5Cpartial+Z_%7B3%7D%7D%3D%5Cltimes_%7B3%7D%5Ccdot+%281-%5Cltimes_%7B3%7D%29&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ltimes_{3}}{&#92;partial Z_{3}}=&#92;ltimes_{3}&#92;cdot (1-&#92;ltimes_{3})\" class=\"latex\" \/><\/pre>\n<p>Finally, the third component of chain rule is again very easy to compute i.e.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+Z_%7B3%7D%7D%7B%5Cpartial+W_%7B5%7D%7D%3D%5Cltimes_%7B1%7D%C2%A0&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial Z_{3}}{&#92;partial W_{5}}=&#92;ltimes_{1}\u00a0\" class=\"latex\" \/><\/pre>\n<p>Since we know,<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=Z_%7B3%7D%3DW_%7B5%7D+%5Ccdot+%5Cltimes_%7B1%7D%C2%A0%2B+W_%7B6%7D+%5Ccdot+%5Cltimes_%7B1%7D%C2%A0+%2B+b_%7B3%7D+&#038;bg=ffffff&#038;fg=000&#038;s=1&#038;c=20201002\" alt=\"Z_{3}=W_{5} &#92;cdot &#92;ltimes_{1}\u00a0+ W_{6} &#92;cdot &#92;ltimes_{1}\u00a0 + b_{3} \" class=\"latex\" \/><\/pre>\n<p>Now, we just multiply these three components of the chain rule and we get the output i.e.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+W_%7B5%7D%7D%3D+%5Cfrac%7B%28%5Cltimes_%7B3%7D-y%29%5Ccdot+%281-%5Cltimes_%7B3%7D%29%5Ccdot%C2%A0+%5Cltimes_%7B3%7D+%5Ccdot+%5Cltimes_%7B1%7D%7D%7B%281-%5Cltimes_%7B3%7D%29%5Ccdot+%5Cltimes_%7B3%7D%7D+%3D+%28%5Cltimes_%7B3%7D-y%29%5Ccdot+%5Cltimes_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial W_{5}}= &#92;frac{(&#92;ltimes_{3}-y)&#92;cdot (1-&#92;ltimes_{3})&#92;cdot\u00a0 &#92;ltimes_{3} &#92;cdot &#92;ltimes_{1}}{(1-&#92;ltimes_{3})&#92;cdot &#92;ltimes_{3}} = (&#92;ltimes_{3}-y)&#92;cdot &#92;ltimes_{1} \" class=\"latex\" \/><\/pre>\n<h2><span style=\"color: #3366ff;\">Chain Rule for W<sub>1 <\/sub>&#8211; Math of Deep Learning<\/span><\/h2>\n<p>The chain rule for the red knob or the additional layer is just an extension of the chain rule of the knob on the top of the tap. This one has a few more components because the water has to travel through more components i.e.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+W_%7B1%7D%7D%3D+%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+%5Cltimes_%7B3%7D%7D+%5Ccdot+%5Cfrac%7B%5Cpartial+%5Cltimes_%7B3%7D%7D%7B%5Cpartial+Z_%7B3%7D%7D%5Ccdot+%5Cfrac%7B%5Cpartial+Z_%7B3%7D%7D%7B%5Cpartial+%5Cltimes_%7B1%7D%7D%5Ccdot+%5Cfrac%7B%5Cpartial+%5Cltimes_%7B1%7D%7D%7B%5Cpartial+Z_%7B1%7D%7D%5Ccdot+%5Cfrac%7B%5Cpartial+Z_%7B1%7D%7D%7B%5Cpartial+W_%7B1%7D%7D+&#038;bg=ffffff&#038;fg=000&#038;s=3&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial W_{1}}= &#92;frac{&#92;partial &#92;ell f }{&#92;partial &#92;ltimes_{3}} &#92;cdot &#92;frac{&#92;partial &#92;ltimes_{3}}{&#92;partial Z_{3}}&#92;cdot &#92;frac{&#92;partial Z_{3}}{&#92;partial &#92;ltimes_{1}}&#92;cdot &#92;frac{&#92;partial &#92;ltimes_{1}}{&#92;partial Z_{1}}&#92;cdot &#92;frac{&#92;partial Z_{1}}{&#92;partial W_{1}} \" class=\"latex\" \/><\/pre>\n<p>The first two components are exactly the same as the knob of the tap i.e. W<sub>5<\/sub>. This makes sense since the water is flowing through the same pipeline towards the end. Hence, we will calculate the third component<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+Z_%7B3%7D%7D%7B%5Cpartial+%5Cltimes_%7B1%7D%7D+%3DW_%7B5%7D&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial Z_{3}}{&#92;partial &#92;ltimes_{1}} =W_{5}\" class=\"latex\" \/><\/pre>\n<p>The fourth component is the derivative of the sigmoid function i.e.\u00a0<strong><a href=\"http:\/\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/09\/Derivative-of-Sigmoid.jpg\" rel=\"noopener\">the derivative of the sigmoid function<\/a><\/strong><\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cltimes_%7B1%7D%7D%7B%5Cpartial+Z_%7B1%7D%7D%3D%5Cltimes_%7B1%7D%5Ccdot+%281-%5Cltimes_%7B1%7D%29&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ltimes_{1}}{&#92;partial Z_{1}}=&#92;ltimes_{1}&#92;cdot (1-&#92;ltimes_{1})\" class=\"latex\" \/><\/pre>\n<p>The fifth and the final component is again easy to calculate.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+Z_%7B1%7D%7D%7B%5Cpartial+W_%7B1%7D%7D%3DX_%7B1%7D&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial Z_{1}}{&#92;partial W_{1}}=X_{1}\" class=\"latex\" \/><\/pre>\n<p>That&#8217;s it. We now multiply these five components to get the results we have already seen for the additional red knob.<\/p>\n<pre><img decoding=\"async\" src=\"https:\/\/s0.wp.com\/latex.php?latex=%5Cfrac%7B%5Cpartial+%5Cell+f+%7D%7B%5Cpartial+W_%7B1%7D%7D%3D+%5Cltimes_%7B3%7D%5Ccdot+%281-%5Cltimes_%7B3%7D%29+%5Ccdot+W_%7B5%7D+%5Ccdot+%5Cltimes_%7B1%7D%5Ccdot+%281-%5Cltimes_%7B1%7D%29+%5Ccdot+X_%7B1%7D+&#038;bg=ffffff&#038;fg=000&#038;s=2&#038;c=20201002\" alt=\"&#92;frac{&#92;partial &#92;ell f }{&#92;partial W_{1}}= &#92;ltimes_{3}&#92;cdot (1-&#92;ltimes_{3}) &#92;cdot W_{5} &#92;cdot &#92;ltimes_{1}&#92;cdot (1-&#92;ltimes_{1}) &#92;cdot X_{1} \" class=\"latex\" \/><\/pre>\n<h4><span style=\"color: #3366ff;\">Sign-off Node<\/span><\/h4>\n<p>This part of the series became a little math heavy. All this, however, will help us a lot when we\u00a0 will build an artificial intelligence to recognize images. See you then.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Welcome back to this series of articles on deep learning and neural networks. In the last part, you learned how training a\u00a0deep learning network is similar to a plumbing job. This time you will learn the math of deep learning. We will continue to use the plumbing analogy to simplify the seemingly complicated math. I<\/p>\n<p><a class=\"excerpt-more blog-excerpt\" href=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/\">Read More&#8230;<\/a><\/p>\n","protected":false},"author":1,"featured_media":11291,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_newsletter_tier_id":0,"jetpack_publicize_message":"","jetpack_is_tweetstorm":false,"jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","enabled":false}}},"categories":[86],"tags":[],"jetpack_publicize_connections":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v17.4 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |\" \/>\n<meta property=\"og:description\" content=\"Welcome back to this series of articles on deep learning and neural networks. In the last part, you learned how training a\u00a0deep learning network is similar to a plumbing job. This time you will learn the math of deep learning. We will continue to use the plumbing analogy to simplify the seemingly complicated math. IRead More...\" \/>\n<meta property=\"og:url\" content=\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/\" \/>\n<meta property=\"og:site_name\" content=\"YOU CANalytics |\" \/>\n<meta property=\"article:author\" content=\"roopam\" \/>\n<meta property=\"article:published_time\" content=\"2018-10-01T04:19:12+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2019-04-21T13:00:41+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&#038;ssl=1\" \/>\n\t<meta property=\"og:image:width\" content=\"796\" \/>\n\t<meta property=\"og:image:height\" content=\"597\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Roopam Upadhyay\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"13 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Organization\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#organization\",\"name\":\"YOU CANalytics\",\"url\":\"https:\/\/ucanalytics.com\/blogs\/\",\"sameAs\":[],\"logo\":{\"@type\":\"ImageObject\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#logo\",\"inLanguage\":\"en-US\",\"url\":\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2015\/11\/YOU-CANalytics-Logo.jpg?fit=607%2C120\",\"contentUrl\":\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2015\/11\/YOU-CANalytics-Logo.jpg?fit=607%2C120\",\"width\":607,\"height\":120,\"caption\":\"YOU CANalytics\"},\"image\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#logo\"}},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#website\",\"url\":\"https:\/\/ucanalytics.com\/blogs\/\",\"name\":\"YOU CANalytics |\",\"description\":\"Explore the Power of Data Science\",\"publisher\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/ucanalytics.com\/blogs\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"ImageObject\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage\",\"inLanguage\":\"en-US\",\"url\":\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1\",\"contentUrl\":\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1\",\"width\":796,\"height\":597},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage\",\"url\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/\",\"name\":\"Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |\",\"isPartOf\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage\"},\"datePublished\":\"2018-10-01T04:19:12+00:00\",\"dateModified\":\"2019-04-21T13:00:41+00:00\",\"breadcrumb\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/ucanalytics.com\/blogs\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Math of Deep Learning Neural Networks &#8211; Simplified (Part 2)\"}]},{\"@type\":\"Article\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage\"},\"author\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#\/schema\/person\/55961a1cea272ecdf290cb387be069b6\"},\"headline\":\"Math of Deep Learning Neural Networks &#8211; Simplified (Part 2)\",\"datePublished\":\"2018-10-01T04:19:12+00:00\",\"dateModified\":\"2019-04-21T13:00:41+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage\"},\"wordCount\":2134,\"commentCount\":6,\"publisher\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#organization\"},\"image\":{\"@id\":\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1\",\"articleSection\":[\"Deep Learning Neural Networks\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#respond\"]}]},{\"@type\":\"Person\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#\/schema\/person\/55961a1cea272ecdf290cb387be069b6\",\"name\":\"Roopam Upadhyay\",\"image\":{\"@type\":\"ImageObject\",\"@id\":\"https:\/\/ucanalytics.com\/blogs\/#personlogo\",\"inLanguage\":\"en-US\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/dd1aa0b0e813f7639800bcfad6a554f1?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/dd1aa0b0e813f7639800bcfad6a554f1?s=96&d=mm&r=g\",\"caption\":\"Roopam Upadhyay\"},\"description\":\"This blog contains my personal views and thoughts on predictive Analytics and big data. - Roopam Upadhyay\",\"sameAs\":[\"roopam\"],\"url\":\"https:\/\/ucanalytics.com\/blogs\/author\/roopam\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/","og_locale":"en_US","og_type":"article","og_title":"Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |","og_description":"Welcome back to this series of articles on deep learning and neural networks. In the last part, you learned how training a\u00a0deep learning network is similar to a plumbing job. This time you will learn the math of deep learning. We will continue to use the plumbing analogy to simplify the seemingly complicated math. IRead More...","og_url":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/","og_site_name":"YOU CANalytics |","article_author":"roopam","article_published_time":"2018-10-01T04:19:12+00:00","article_modified_time":"2019-04-21T13:00:41+00:00","og_image":[{"width":796,"height":597,"url":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1","type":"image\/jpeg"}],"twitter_misc":{"Written by":"Roopam Upadhyay","Est. reading time":"13 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Organization","@id":"https:\/\/ucanalytics.com\/blogs\/#organization","name":"YOU CANalytics","url":"https:\/\/ucanalytics.com\/blogs\/","sameAs":[],"logo":{"@type":"ImageObject","@id":"https:\/\/ucanalytics.com\/blogs\/#logo","inLanguage":"en-US","url":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2015\/11\/YOU-CANalytics-Logo.jpg?fit=607%2C120","contentUrl":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2015\/11\/YOU-CANalytics-Logo.jpg?fit=607%2C120","width":607,"height":120,"caption":"YOU CANalytics"},"image":{"@id":"https:\/\/ucanalytics.com\/blogs\/#logo"}},{"@type":"WebSite","@id":"https:\/\/ucanalytics.com\/blogs\/#website","url":"https:\/\/ucanalytics.com\/blogs\/","name":"YOU CANalytics |","description":"Explore the Power of Data Science","publisher":{"@id":"https:\/\/ucanalytics.com\/blogs\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/ucanalytics.com\/blogs\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"ImageObject","@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage","inLanguage":"en-US","url":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1","contentUrl":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1","width":796,"height":597},{"@type":"WebPage","@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage","url":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/","name":"Math of Deep Learning Neural Networks - Simplified (Part 2) &ndash; YOU CANalytics |","isPartOf":{"@id":"https:\/\/ucanalytics.com\/blogs\/#website"},"primaryImageOfPage":{"@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage"},"datePublished":"2018-10-01T04:19:12+00:00","dateModified":"2019-04-21T13:00:41+00:00","breadcrumb":{"@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/ucanalytics.com\/blogs\/"},{"@type":"ListItem","position":2,"name":"Math of Deep Learning Neural Networks &#8211; Simplified (Part 2)"}]},{"@type":"Article","@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#article","isPartOf":{"@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage"},"author":{"@id":"https:\/\/ucanalytics.com\/blogs\/#\/schema\/person\/55961a1cea272ecdf290cb387be069b6"},"headline":"Math of Deep Learning Neural Networks &#8211; Simplified (Part 2)","datePublished":"2018-10-01T04:19:12+00:00","dateModified":"2019-04-21T13:00:41+00:00","mainEntityOfPage":{"@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#webpage"},"wordCount":2134,"commentCount":6,"publisher":{"@id":"https:\/\/ucanalytics.com\/blogs\/#organization"},"image":{"@id":"https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#primaryimage"},"thumbnailUrl":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1","articleSection":["Deep Learning Neural Networks"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/ucanalytics.com\/blogs\/math-of-deep-learning-neural-networks-simplified-part-2\/#respond"]}]},{"@type":"Person","@id":"https:\/\/ucanalytics.com\/blogs\/#\/schema\/person\/55961a1cea272ecdf290cb387be069b6","name":"Roopam Upadhyay","image":{"@type":"ImageObject","@id":"https:\/\/ucanalytics.com\/blogs\/#personlogo","inLanguage":"en-US","url":"https:\/\/secure.gravatar.com\/avatar\/dd1aa0b0e813f7639800bcfad6a554f1?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/dd1aa0b0e813f7639800bcfad6a554f1?s=96&d=mm&r=g","caption":"Roopam Upadhyay"},"description":"This blog contains my personal views and thoughts on predictive Analytics and big data. - Roopam Upadhyay","sameAs":["roopam"],"url":"https:\/\/ucanalytics.com\/blogs\/author\/roopam\/"}]}},"jetpack_featured_media_url":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-and-Plumbing-Job.jpg?fit=796%2C597&ssl=1","jetpack_sharing_enabled":true,"jetpack_shortlink":"https:\/\/wp.me\/p3L0jT-2W6","jetpack-related-posts":[{"id":11156,"url":"https:\/\/ucanalytics.com\/blogs\/deep-learning-and-neural-networks-simplified-part-1\/","url_meta":{"origin":11290,"position":0},"title":"Deep Learning and Neural Networks \u2013 Simplified (Part 1)","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"The entire field of artificial intelligence, in the last few years, is built upon deep learning or deep neural networks. Notably, Apple's Siri, Google-DeepMinds' AlphaGo, or the self-driving mechanism in Tesla cars are all based on deep learning. Here, my goal is to make deep learning neural networks\u00a0much more accessible\u2026","rel":"","context":"In &quot;Deep Learning Neural Networks&quot;","block_context":{"text":"Deep Learning Neural Networks","link":"https:\/\/ucanalytics.com\/blogs\/category\/deep-learning-neural-networks\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Front.jpg?fit=960%2C592&ssl=1&resize=350%2C200","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Front.jpg?fit=960%2C592&ssl=1&resize=350%2C200 1x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Front.jpg?fit=960%2C592&ssl=1&resize=525%2C300 1.5x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/08\/Deep-Learning-Neural-Networks-Front.jpg?fit=960%2C592&ssl=1&resize=700%2C400 2x"},"classes":[]},{"id":6808,"url":"https:\/\/ucanalytics.com\/blogs\/learning-resources-artificial-intelligence-cognitive-computing-deep-learning-neural-networks\/","url_meta":{"origin":11290,"position":1},"title":"Learning Resources : Artificial Intelligence, Cognitive Computing, Deep Learning, &#038; Neural Networks","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"This article is an effort to make you into a \"semi-expert\" in artificial intelligence, cognitive computing, deep learning and neural networks from scratch. Here I will share a few cool learning resources for these topics. These resources include documentaries, TED talks, online lecture videos, and books. There are several videos\u2026","rel":"","context":"In &quot;Analytics Book Club&quot;","block_context":{"text":"Analytics Book Club","link":"https:\/\/ucanalytics.com\/blogs\/category\/analytics-book-club\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2015\/11\/Alice-in-Wonderland-Copy.jpg?fit=433%2C274&ssl=1&resize=350%2C200","width":350,"height":200},"classes":[]},{"id":11578,"url":"https:\/\/ucanalytics.com\/blogs\/deep-learning-models-simplified-part-3\/","url_meta":{"origin":11290,"position":2},"title":"Deep Learning Models Simplified (Part 3)","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"Facebook was a major sensation and a source of great amusement in a British\u00a0country house in the early 20th century. It was such a big hit that it got a special mention in a newspaper published in the year 1902. Facebook, then, of course, had a completely different meaning than\u2026","rel":"","context":"In &quot;Deep Learning Neural Networks&quot;","block_context":{"text":"Deep Learning Neural Networks","link":"https:\/\/ucanalytics.com\/blogs\/category\/deep-learning-neural-networks\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Popeye-Deep-Learning.jpg?fit=960%2C686&ssl=1&resize=350%2C200","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Popeye-Deep-Learning.jpg?fit=960%2C686&ssl=1&resize=350%2C200 1x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Popeye-Deep-Learning.jpg?fit=960%2C686&ssl=1&resize=525%2C300 1.5x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Popeye-Deep-Learning.jpg?fit=960%2C686&ssl=1&resize=700%2C400 2x"},"classes":[]},{"id":4054,"url":"https:\/\/ucanalytics.com\/blogs\/artificial-neural-networks-retail-case-study-example-part-8\/","url_meta":{"origin":11290,"position":3},"title":"Artificial Neural Networks, Made Easy &#8211; Retail Case Study Example (Part 8)","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"Welcome back to\u00a0our retail case study example\u00a0for campaign and marketing analytics. So far in this case study, we were working on a\u00a0classification problem to identify customers with a higher likelihood to purchase products from\u00a0the campaign catalogues. In the last article on\u00a0model selection, we noticed that artificial neural networks performed better\u2026","rel":"","context":"In &quot;Marketing Analytics&quot;","block_context":{"text":"Marketing Analytics","link":"https:\/\/ucanalytics.com\/blogs\/category\/marketing-analytics\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/Artificial-Neural-Networks.jpg?fit=346%2C336&ssl=1&resize=350%2C200","width":350,"height":200},"classes":[]},{"id":11583,"url":"https:\/\/ucanalytics.com\/blogs\/convolutional-neural-networks-cnn-simplified-part-4\/","url_meta":{"origin":11290,"position":4},"title":"Convolutional Neural Networks (CNN) Simplified (Part 4)","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"Welcome back to the deep learning example to build an OCR application. The idea of this simple application is to identify numbers in an image of written text. In the last part, we used three different models and got the following accuracy for identification of the test images: Model 1\u2026","rel":"","context":"In &quot;Deep Learning Neural Networks&quot;","block_context":{"text":"Deep Learning Neural Networks","link":"https:\/\/ucanalytics.com\/blogs\/category\/deep-learning-neural-networks\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Convolutional-Neural-Networks-CNN-Filters-Polaroid-Glasses-by-Roopam.jpg?fit=775%2C581&ssl=1&resize=350%2C200","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Convolutional-Neural-Networks-CNN-Filters-Polaroid-Glasses-by-Roopam.jpg?fit=775%2C581&ssl=1&resize=350%2C200 1x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Convolutional-Neural-Networks-CNN-Filters-Polaroid-Glasses-by-Roopam.jpg?fit=775%2C581&ssl=1&resize=525%2C300 1.5x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2018\/10\/Convolutional-Neural-Networks-CNN-Filters-Polaroid-Glasses-by-Roopam.jpg?fit=775%2C581&ssl=1&resize=700%2C400 2x"},"classes":[]},{"id":3973,"url":"https:\/\/ucanalytics.com\/blogs\/model-selection-retail-case-study-example-part-7\/","url_meta":{"origin":11290,"position":5},"title":"Model Selection &#8211; Retail Case Study Example (Part 7)","author":"Roopam Upadhyay","date":false,"format":false,"excerpt":"Model Selection This is a continuation of our retail case study example for campaign and marketing analytics. In the previous two parts, we discussed a couple of decision tree algorithms (CART and C4.5)\u00a0for classification. Recall a previous case study example on\u00a0banking and risk management where we discussed logistic regression\u00a0which is\u2026","rel":"","context":"In &quot;Marketing Analytics&quot;","block_context":{"text":"Marketing Analytics","link":"https:\/\/ucanalytics.com\/blogs\/category\/marketing-analytics\/"},"img":{"alt_text":"","src":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/photo.jpg?fit=1200%2C1029&ssl=1&resize=350%2C200","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/photo.jpg?fit=1200%2C1029&ssl=1&resize=350%2C200 1x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/photo.jpg?fit=1200%2C1029&ssl=1&resize=525%2C300 1.5x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/photo.jpg?fit=1200%2C1029&ssl=1&resize=700%2C400 2x, https:\/\/i0.wp.com\/ucanalytics.com\/blogs\/wp-content\/uploads\/2014\/09\/photo.jpg?fit=1200%2C1029&ssl=1&resize=1050%2C600 3x"},"classes":[]}],"_links":{"self":[{"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/posts\/11290"}],"collection":[{"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/comments?post=11290"}],"version-history":[{"count":0,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/posts\/11290\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/media\/11291"}],"wp:attachment":[{"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/media?parent=11290"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/categories?post=11290"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/ucanalytics.com\/blogs\/wp-json\/wp\/v2\/tags?post=11290"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}