Skip to content

Commit 6cf5b46

Browse files
committed
update project 2
1 parent 8e604f1 commit 6cf5b46

File tree

9 files changed

+88
-683
lines changed

9 files changed

+88
-683
lines changed

doc/Projects/2025/Project2/html/._Project2-bs000.html

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -243,14 +243,14 @@ <h2 id="classification-and-regression-writing-our-own-neural-network-code" class
243243
<p>The data sets that we propose here are (the default sets)</p>
244244

245245
<ul>
246-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
247-
<ol type="a"></li>
248-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249-
</ol>
250-
<li> Classification.
251-
<ol type="a"></li>
252-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253-
</ol>
246+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
247+
<ul>
248+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249+
</ul>
250+
<li> Classification.</li>
251+
<ul>
252+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253+
</ul>
254254
</ul>
255255
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
256256
<h3 id="part-a-analytical-warm-up" class="anchor">Part a): Analytical warm-up </h3>
@@ -260,10 +260,10 @@ <h3 id="part-a-analytical-warm-up" class="anchor">Part a): Analytical warm-up </
260260
</p>
261261
<ol>
262262
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
263-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
263+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
264264
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
265265
</ol>
266-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
266+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
267267

268268
<p>We will test three activation functions for our neural network setup, these are the </p>
269269
<ol>
@@ -330,7 +330,7 @@ <h3 id="part-b-writing-your-own-neural-network-code" class="anchor">Part b): Wri
330330
<p>Comment your results and give a critical discussion of the results
331331
obtained with the OLS code from project 1 and your own neural network
332332
code. Make an analysis of the learning rates employed to find the
333-
optimal MSE and \( R2 \) scores. Test both stochastic gradient descent
333+
optimal MSE score. Test both stochastic gradient descent
334334
with RMSprop and ADAM and plain gradient descent with different
335335
learning rates.
336336
</p>

doc/Projects/2025/Project2/html/Project2-bs.html

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -243,14 +243,14 @@ <h2 id="classification-and-regression-writing-our-own-neural-network-code" class
243243
<p>The data sets that we propose here are (the default sets)</p>
244244

245245
<ul>
246-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
247-
<ol type="a"></li>
248-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249-
</ol>
250-
<li> Classification.
251-
<ol type="a"></li>
252-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253-
</ol>
246+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
247+
<ul>
248+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_self"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
249+
</ul>
250+
<li> Classification.</li>
251+
<ul>
252+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_self"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
253+
</ul>
254254
</ul>
255255
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
256256
<h3 id="part-a-analytical-warm-up" class="anchor">Part a): Analytical warm-up </h3>
@@ -260,10 +260,10 @@ <h3 id="part-a-analytical-warm-up" class="anchor">Part a): Analytical warm-up </
260260
</p>
261261
<ol>
262262
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
263-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
263+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
264264
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
265265
</ol>
266-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
266+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
267267

268268
<p>We will test three activation functions for our neural network setup, these are the </p>
269269
<ol>
@@ -330,7 +330,7 @@ <h3 id="part-b-writing-your-own-neural-network-code" class="anchor">Part b): Wri
330330
<p>Comment your results and give a critical discussion of the results
331331
obtained with the OLS code from project 1 and your own neural network
332332
code. Make an analysis of the learning rates employed to find the
333-
optimal MSE and \( R2 \) scores. Test both stochastic gradient descent
333+
optimal MSE score. Test both stochastic gradient descent
334334
with RMSprop and ADAM and plain gradient descent with different
335335
learning rates.
336336
</p>

doc/Projects/2025/Project2/html/Project2.html

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -278,14 +278,14 @@ <h2 id="classification-and-regression-writing-our-own-neural-network-code">Class
278278
<p>The data sets that we propose here are (the default sets)</p>
279279

280280
<ul>
281-
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be
282-
<ol type="a"></li>
283-
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. You should however feel free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_blank"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
284-
</ol>
285-
<li> Classification.
286-
<ol type="a"></li>
287-
<li> We will consider the multiclass classification problem given by the full MNIST data set. The one included in <b>scikit-learn</b> is reduced data. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_blank"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
288-
</ol>
281+
<li> Regression (fitting a continuous function). In this part you will need to bring back your results from project 1 and compare these with what you get from your Neural Network code to be developed here. The data sets could be</li>
282+
<ul>
283+
<li> The simple one-dimensional function Runge function from project 1, that is \( f(x) = \frac{1}{1+25x^2} \). We recommend using a simpler function when developing your neural network code for regression problems. Feel however free to discuss and study other functions, such as the the two-dimensional Runge function \( f(x,y)=\left[(10x - 5)^2 + (10y - 5)^2 + 1 \right]^{-1} \), or even more complicated two-dimensional functions (see the supplementary material of <a href="https://www.nature.com/articles/s41467-025-61362-4" target="_blank"><tt>https://www.nature.com/articles/s41467-025-61362-4</tt></a> for an extensive list of two-dimensional functions).</li>
284+
</ul>
285+
<li> Classification.</li>
286+
<ul>
287+
<li> We will consider a multiclass classification problem given by the full MNIST data set. The full data set is at <a href="https://www.kaggle.com/datasets/hojjatk/mnist-dataset" target="_blank"><tt>https://www.kaggle.com/datasets/hojjatk/mnist-dataset</tt></a>.</li>
288+
</ul>
289289
</ul>
290290
<p>We will start with a regression problem and we will reuse our codes on gradient descent methods from project 1.</p>
291291
<h3 id="part-a-analytical-warm-up">Part a): Analytical warm-up </h3>
@@ -295,10 +295,10 @@ <h3 id="part-a-analytical-warm-up">Part a): Analytical warm-up </h3>
295295
</p>
296296
<ol>
297297
<li> The mean-squared error (MSE) with and without the \( L_1 \) and \( L_2 \) norms (regression problems)</li>
298-
<li> The binary cross entropy (aka log loss) for classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
298+
<li> The binary cross entropy (aka log loss) for binary classification problems with and without \( L_1 \) and \( L_2 \) norms</li>
299299
<li> The multiclass cross entropy cost/loss function (aka Softmax cross entropy or just Softmax loss function)</li>
300300
</ol>
301-
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms.</p>
301+
<p>Set up these three cost/loss functions and their respective derivatives and explain the various terms. In this project you will however only use the MSE and the Softmax cross entropy.</p>
302302

303303
<p>We will test three activation functions for our neural network setup, these are the </p>
304304
<ol>
@@ -365,7 +365,7 @@ <h3 id="part-b-writing-your-own-neural-network-code">Part b): Writing your own N
365365
<p>Comment your results and give a critical discussion of the results
366366
obtained with the OLS code from project 1 and your own neural network
367367
code. Make an analysis of the learning rates employed to find the
368-
optimal MSE and \( R2 \) scores. Test both stochastic gradient descent
368+
optimal MSE score. Test both stochastic gradient descent
369369
with RMSprop and ADAM and plain gradient descent with different
370370
learning rates.
371371
</p>

0 commit comments

Comments
 (0)