Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

index errors with large datasets? #38

Open
whoburg opened this issue May 16, 2016 · 0 comments
Open

index errors with large datasets? #38

whoburg opened this issue May 16, 2016 · 0 comments
Labels

Comments

@whoburg
Copy link
Collaborator

whoburg commented May 16, 2016

reported by tony tao:

"on the surface it looks like a data size issue ("MemoryError" and "Iterator too large" errors) but when I truncate the data set to something that already worked before, it returns indexing errors which makes me believe it's something in GPfit, but I can't figure out what it is. "

"Actually (as usual, problem is solved after calling mayday) I may have figured it out and now I have a Cd model as well.

The training input data set is around 40,000 data points over 7 dimensions (originally 80,000), so it takes around 16 GB of memory to build the model, which explains the memory error running in Python(x,y). Running it in Ubuntu and deleting about half the training data seems to have fixed it.

The index error is caused by line 73 of the max_affine_init.py script where if the while loop isn't fulfilled by the end of the dataset, it calls the next index location which is out of bounds. "

@pgkirsch pgkirsch added the error label Apr 22, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants