Design Tradeoffs In A Hardware Implementation Of The K-Means Clustering Algorithm

SAM 2000: PROCEEDINGS OF THE 2000 IEEE SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP(2000)

引用 29|浏览5
暂无评分
摘要
Hyperspectral imagery provides exquisitely detailed information, but poses a serious challenge to the image analyst Massive quantities of data must be reduced in a way that identifies useful features, Pixel clustering algorithms provide one approach: each pixel is assigned to a class based on the spectral similarity of that pixel to other members of the class. An image where each pixel is represented by its class provides a single picture (not hundreds of channels) for the analyst to interpret,Opportunities for fine-grain parallelism in this computationally intensive calculation motivate the consideration of specialized hardware. We are developing an FPGA-based implementation of the k-means clustering algorithm, because FPGAs can provide considerable speedup yet are sufficiently flexible to permit the testing of variants and tradeoffs.Assigning pixels to classes requires a definition of spectral similarity. The standard choice is the Euclidean distance (sum of the squared differences between the pixel and the cluster center taken on a coordinate by coordinate basis). We consider two alternative distances: the Manhattan (sum of the absolute coordinate differences) and the Mar (maximum of the absolute differences). The Euclidean distance is optimal in the sense of minimizing within-class variance, but requires computing many squares. The Manhattan and Max distances eliminate this squaring, and furthermore reduce the number of bits needed to store intermediate calculations.We conducted two sets of experiments to assess the tradeoff between quality of clustering and efficiency of hardware implementation. The first considered an idealized setup for assessing mis-classification rates and their effect on within-class variance. The second evaluated within-class variance for real 224-channel AVIRIS data sets. Both experiments confirmed that the highest quality clusters are achieved using the Euclidean metric, and quantified the difference. In comparing the faster, alternative distances, we found with the AVIRIS data that the results depended on which image was used; however, the idealized experiment showed a clear preference for the Manhattan metric.
更多
查看译文
关键词
hyperspectral imagery,hyperspectral sensors,parallel algorithms,k means clustering algorithm,concurrent computing,hyperspectral imaging,information analysis,parallel processing,field programmable gate arrays,pixel,image analysis,euclidean metric,clustering algorithms,euclidean distance,manhattan distance,hardware,fpga,k means clustering,image classification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要