Today, petascale distributed memory systems perform large-scale simulations and generate massive amounts of data in a distributed fashion at unprecedented rates. When clustering data, traditional analysis methods may require the comparison of single records with each other in an iterative process, and therefore involve moving data across system nodes. When both the data and the number of nodes increase, clustering methods can increase pressure on storage and bandwidth. Thus, the methods become inefficient and do not scale. New methodologies are needed to analyze data when it is distributed across nodes of large distributed memory systems.
When analyzing structural biology datasets, we focus on specific properties of the data records such as the molecular geometry or the location of a molecule in a docking pocket. In this talk we propose a methodology that allows the scalable analysis for large datasets composed of millions of individual structural biology records in a distributed manner on large distributed memory systems. The methodology is based on two general steps. The first step extracts concise properties or features of each data record and represents them as metadata in parallel. The second step performs the clustering on the extracted properties using machine-learning techniques. We apply the methodology to two different computational structural biology datasets to identify geometrical features. Our results show that our approach enables scalable clustering analyses for large-scale computational structural biology datasets on large distributed memory systems, and that our method achieves better accuracy compared to traditional analysis approaches.