Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DOC: Adds Inigo's week 5 short blogpost #45

Merged
merged 3 commits into from
Jul 4, 2024

Conversation

itellaetxe
Copy link
Contributor

Adds what I did on week 5, mainly vacation time.

Copy link
Contributor

github-actions bot commented Jul 1, 2024

🪓 PR closed, deleted preview at https://github.com/dipy/preview-html/tree/main/dipy.org/pull/45/

posts/2024/2024_06_28_Inigo_week_5.rst Outdated Show resolved Hide resolved
posts/2024/2024_06_28_Inigo_week_5.rst Outdated Show resolved Hide resolved
posts/2024/2024_06_28_Inigo_week_5.rst Outdated Show resolved Hide resolved
posts/2024/2024_06_28_Inigo_week_5.rst Show resolved Hide resolved
Copy link
Member

@skoudoro skoudoro left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi @itellaetxe,

From my side, it looks good, just waiting that you address @jhlegarreta comments.

Hi everyone! This week I have been on vacation, so I have not been able to work on the project as much as the previous weeks. However, I have been thinking about the next steps to take and I have decided to start with the conditional AutoEncoder. I have been reading some papers and I have found some interesting ideas that would be nice to implement.

While stuck in the Munich airport, I started to write some code for this (the weather was not very good and my flight was delayed, so I lost my connecting flight and I had to sleep in the airport). I found an implementation of a regression variational AutoEncoder `in this paper <https://doi.org/10.1007/978-3-030-32245-8_91>`_, where the authors implement a way to manipulate the latent space so that the samples (streamlines) that get projected into it are organized along a desired scalar parameter.
I thought this could be a good starting point for my conditional AutoEncoder because it basically provides a way to sample from the latent space in a controlled manner, where you can select the age of the streamlines you want to generate. Also, having the variational part makes the latent space more regularized, so our model is more resilient against overfitting, which is the main flaw of ordinary AutoEncoders.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also, I am all that comfortable saying that overfitting is an inherent property to AEs. Do you have a scientific reference that supports this?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thnks for pointing this out. To be precise, I should rephrase this to "...overfitting, what could happen when non-regularized (vanilla) AutoEncoders learn the identity function, which leads to lack of generalizability outside the test set."

This is explained in this NeurIPS 2020 paper (PDF)
image

In conclusion, yes, saying that "AEs main flaw is (natural) overfitting" is too much. However, saying that they might overfit towards the identity function, which is the effect I was aware of at the time I wrote the blog, is more rigurous.

What do you think @jhlegarreta ?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the reference.

"which leads to lack of generalizability outside the test set." outside the training set ?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

True, sorry. Yes, I meant "outside the training set"

@skoudoro skoudoro merged commit 65c3b20 into dipy:master Jul 4, 2024
3 checks passed
github-actions bot added a commit that referenced this pull request Jul 4, 2024
* DOC: Adds Inigo's week 5 short blogpost

* FIX: Grammar issues and paragraph spacing

* FIX: Corrects overfitting comment 65c3b20
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants