You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The reason it takes so long to run is not because of image being too large, it is because of the leafsize chosen is too close to the coreset_size
if you fix leafsize issue, the algorithm runs much faster even for downsampling = 1
However, the plot (middle one) does not get generated, it is because of the markersize, s=np.exp(2.0 * coreset_size * herding_weights).reshape(1, -1)
If you downsample by a factor of n, the number of datapoints is reduced by factor of n^2, this means if you choose downsampling size = 8, you get an error (you also get error when you choose n = 2, I don't know why), this is because, at this point you try to obtain a coreset of size (8000/n), from original data of size (36,000/n**2)
Suggested fix:
change the interpolation (using resize area interpolation)
wherever we were dividing by downsample_scale, we divide by downsample_scale ** 2
replace the marker size by 1 (this is only temporary fix, and investigate why this problem exists)
change leafsize to 16,000 or 24,000 from 10,000 (this speeds up the code significantly)
How can we reproduce the issue?
Change leafsize to 16,000 or 24,000 from 10,000 to fix the issue associated with the first point and run the David example with downsample_size = 1 (default)
However, if you now replace the markersize s to 1, it runs perfectly fine
Python version
3.12
Package version
0.2.1
Operating system
Windows
Other packages
No response
Relevant log output
No response
The text was updated successfully, but these errors were encountered:
What's the problem?
I believe that:
The reason it takes so long to run is not because of image being too large, it is because of the
leafsize
chosen is too close to thecoreset_size
if you fix leafsize issue, the algorithm runs much faster even for downsampling = 1
However, the plot (middle one) does not get generated, it is because of the markersize, s=np.exp(2.0 * coreset_size * herding_weights).reshape(1, -1)
If you downsample by a factor of n, the number of datapoints is reduced by factor of n^2, this means if you choose downsampling size = 8, you get an error (you also get error when you choose n = 2, I don't know why), this is because, at this point you try to obtain a coreset of size (8000/n), from original data of size (36,000/n**2)
Suggested fix:
downsample_scale
, we divide bydownsample_scale ** 2
leafsize
to 16,000 or 24,000 from 10,000 (this speeds up the code significantly)How can we reproduce the issue?
Change
leafsize
to16,000
or24,000
from10,000
to fix the issue associated with the first point and run the David example withdownsample_size
= 1 (default)However, if you now replace the markersize
s
to 1, it runs perfectly finePython version
3.12
Package version
0.2.1
Operating system
Windows
Other packages
No response
Relevant log output
No response
The text was updated successfully, but these errors were encountered: