I am focusing on machine learning. Currently i am improving naive bayes text classifier. I am not using library but on my own coding from scratch.
Having enthusiast on research to provide better and original algorithm. One of them is able to reduce dataset more compact so that we can have dataset from big number of rows into very small number of rows compared to number of rows from query.
Suppose you have dataset with 10 thousand rows for specified query with 5 thousand rows, then using dataset optimizer, we can reduce dataset into 5 thousand rows or even (very) less. This i may call sufficient dataset.
If you want to try it, you can send your dataset at least 10 thousand rows and your query (must be relevant to dataset)
Please i can’t tackle any of your request. I’ll pick one of them and send it back to you for your review.
This must give a valuable dataset which maintain accuracy or improve accuracy.
Thanks in advance