Accelerate runtime when data set is large
When data set size goes to millions of rows and hundreds of features, it takes hours to run. Could there be ways to shorten the computing time?
Perhaps you might try our other package https://github.com/almost-matching-exactly/variable_imp_matching ? This one might scale better. Or you could try our database option.
On Nov 8, 2023, at 10:22 AM, wtc100 @.***> wrote:
When data set size goes to millions of rows and hundreds of features, it takes hours to run. Could there be ways to shorten the computing time? — Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you are subscribed to this thread.Message ID: @.***>
I want to add the link to the database version (the flame_db folder): https://github.com/almost-matching-exactly/DAME-FLAME-Python-Package/tree/2d941bcfa76d7bcd33d58cbf4657202e62cc5b0c
and its documentation: https://github.com/almost-matching-exactly/DAME-FLAME-Python-Package?tab=readme-ov-file#a-tutorial-to-flame-database-version