{"id":9376,"date":"2013-05-15T07:13:16","date_gmt":"2013-05-15T04:13:16","guid":{"rendered":"http:\/\/hgpu.org\/?p=9376"},"modified":"2013-05-15T07:13:16","modified_gmt":"2013-05-15T04:13:16","slug":"approximative-inference-for-multivariate-functional-data-on-massively-parallel-processors","status":"publish","type":"post","link":"https:\/\/hgpu.org\/?p=9376","title":{"rendered":"Approximative inference for multivariate functional data on massively parallel processors"},"content":{"rendered":"<p>With continually increasing data sizes, the relevance of the big n problem of classical likelihood approaches is greater than ever. This paper considers functional data, and presents operator approximations, where observations are embedded in function space, and likelihood calculations are carried out in the functional domain. The resulting approximated problems are naturally parallel and can be solved in linear time. An extremely efficient GPU implementation is presented, and the proposed methods are illustrated by conducting a classical statistical analysis of a dataset of 2D chromatograms consisting of more than 140 million spatially correlated observation points.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>With continually increasing data sizes, the relevance of the big n problem of classical likelihood approaches is greater than ever. This paper considers functional data, and presents operator approximations, where observations are embedded in function space, and likelihood calculations are carried out in the functional domain. The resulting approximated problems are naturally parallel and can [&hellip;]<\/p>\n","protected":false},"author":351,"featured_media":0,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[89,157,3],"tags":[14,1796,20,378],"class_list":["post-9376","post","type-post","status-publish","format-standard","hentry","category-nvidia-cuda","category-mathematics","category-paper","tag-cuda","tag-mathematics","tag-nvidia","tag-tesla-c2050"],"views":1961,"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts\/9376","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/users\/351"}],"replies":[{"embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=9376"}],"version-history":[{"count":0,"href":"https:\/\/hgpu.org\/index.php?rest_route=\/wp\/v2\/posts\/9376\/revisions"}],"wp:attachment":[{"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=9376"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=9376"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/hgpu.org\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=9376"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}