You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: doc/Projects/2025/Project2/html/._Project2-bs000.html
+11-11Lines changed: 11 additions & 11 deletions
Original file line number
Diff line number
Diff line change
@@ -243,14 +243,14 @@ <h2 id="classification-and-regression-writing-our-own-neural-network-code" class
243
243
<p>The data sets that we propose here are (the default sets)</p>
244
244
245
245
<ul>
246
-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
247
-
<oltype="a"></li>
248
-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249
-
</ol>
250
-
<li> Classification.
251
-
<oltype="a"></li>
252
-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253
-
</ol>
246
+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
247
+
<ul>
248
+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249
+
</ul>
250
+
<li> Classification.</li>
251
+
<ul>
252
+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253
+
</ul>
254
254
</ul>
255
255
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
263
-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
263
+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
264
264
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
265
265
</ol>
266
-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
266
+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
267
267
268
268
<p>We will test three activation functions for our neural network setup, these are the </p>
Copy file name to clipboardExpand all lines: doc/Projects/2025/Project2/html/Project2-bs.html
+11-11Lines changed: 11 additions & 11 deletions
Original file line number
Diff line number
Diff line change
@@ -243,14 +243,14 @@ <h2 id="classification-and-regression-writing-our-own-neural-network-code" class
243
243
<p>The data sets that we propose here are (the default sets)</p>
244
244
245
245
<ul>
246
-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
247
-
<oltype="a"></li>
248
-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249
-
</ol>
250
-
<li> Classification.
251
-
<oltype="a"></li>
252
-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253
-
</ol>
246
+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
247
+
<ul>
248
+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249
+
</ul>
250
+
<li> Classification.</li>
251
+
<ul>
252
+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253
+
</ul>
254
254
</ul>
255
255
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
263
-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
263
+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
264
264
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
265
265
</ol>
266
-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
266
+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
267
267
268
268
<p>We will test three activation functions for our neural network setup, these are the </p>
<p>The data sets that we propose here are (the default sets)</p>
279
279
280
280
<ul>
281
-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
282
-
<oltype="a"></li>
283
-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_blank"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
284
-
</ol>
285
-
<li> Classification.
286
-
<oltype="a"></li>
287
-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_blank"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
288
-
</ol>
281
+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
282
+
<ul>
283
+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <ahref="https://www.nature.com/articles/s41467-025-61362-4" target="_blank"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
284
+
</ul>
285
+
<li> Classification.</li>
286
+
<ul>
287
+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <ahref="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_blank"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
288
+
</ul>
289
289
</ul>
290
290
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
298
-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
298
+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
299
299
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
300
300
</ol>
301
-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
301
+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
302
302
303
303
<p>We will test three activation functions for our neural network setup, these are the </p>
304
304
<ol>
@@ -365,7 +365,7 @@ <h3 id="part-b-writing-your-own-neural-network-code">Part b): Writing your own N
365
365
<p>Comment your results and give a critical discussion of the results
366
366
obtained with the OLS code from project 1 and your own neural network
367
367
code. Make an analysis of the learning rates employed to find the
368
-
optimal MSE and \( R2 \) scores. Test both stochastic gradient descent
368
+
optimal MSE score. Test both stochastic gradient descent
369
369
with RMSprop and ADAM and plain gradient descent with different
0 commit comments