A modified method of calculating High Dimensional Model Representation (HDMR) Terms for parallelization with MPI and CUDA


Kanal M. E., Demiralp M.

JOURNAL OF SUPERCOMPUTING, cilt.62, sa.1, ss.199-213, 2012 (SCI-Expanded) identifier identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 62 Sayı: 1
  • Basım Tarihi: 2012
  • Doi Numarası: 10.1007/s11227-011-0695-0
  • Dergi Adı: JOURNAL OF SUPERCOMPUTING
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus
  • Sayfa Sayıları: ss.199-213
  • İstanbul Teknik Üniversitesi Adresli: Evet

Özet

If the values of a multivariate function f(x (1),x (2),aEuro broken vertical bar,x (N) ) are given at only a finite number of points in the space of its arguments and an interpolation which employs continuous functions is considered standard multivariate routines may become cumbersome as the dimensionality grows. This urges us to develop a divide-and-conquer algorithm which approximates the function. The given multivariate data are partitioned into low-variate data. This approach is called High Dimensional Model Representation (HDMR). However, the method in its current form is not applicable to problems having huge volumes of data. With the increasing dimension number and the number of the corresponding nodes, the volume of data in question reaches such a high level that it is beyond the capacity of any individual PC because huge volume of data requires much higher RAM capacity. Another aspect is that the structure of equalities used in the calculation of HDMR terms varies according to the dimension number of the problem. The number of loops in the algorithm increases with the increasing dimension number. In this work, as a first step, the equations used are modified in such a way that their structure does not depend on the dimension number. With the newly obtained equalities, the method becomes appropriate for parallelization. Due to the parallelization, the RAM problem arising from problems with high volume of data is solved. Finally, the performance of the parallelized method is analyzed.