General Description
Our goal is to create algorithms to detect duplicated or unnecessary data in large data sets. This can easily occur when a scientist collects data and when this happens the data becomes more difficult to deal with and wastes the finite resources that the computer has at it's disposal. This can also lead to a slow down among other issues. Once we detect the unnecessary or duplicated data we would like to find a way to dispose of this data.
Comments
Post a Comment