TY - JOUR
T1 - Parallel generation of architecture on the GPU
AU - Steinberger, Markus
AU - Kenzel, Michael
AU - Kainz, Bernhard K.
AU - Müller, Jörg
AU - Wonka, Peter
AU - Schmalstieg, Dieter
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledgements: This research was funded by the Austrian Science Fund (FWF): P23329.
PY - 2014/6/1
Y1 - 2014/6/1
N2 - In this paper, we present a novel approach for the parallel evaluation of procedural shape grammars on the graphics processing unit (GPU). Unlike previous approaches that are either limited in the kind of shapes they allow, the amount of parallelism they can take advantage of, or both, our method supports state of the art procedural modeling including stochasticity and context-sensitivity. To increase parallelism, we explicitly express independence in the grammar, reduce inter-rule dependencies required for context-sensitive evaluation, and introduce intra-rule parallelism. Our rule scheduling scheme avoids unnecessary back and forth between CPU and GPU and reduces round trips to slow global memory by dynamically grouping rules in on-chip shared memory. Our GPU shape grammar implementation is multiple orders of magnitude faster than the standard in CPU-based rule evaluation, while offering equal expressive power. In comparison to the state of the art in GPU shape grammar derivation, our approach is nearly 50 times faster, while adding support for geometric context-sensitivity. © 2014 The Author(s) Computer Graphics Forum © 2014 The Eurographics Association and John Wiley & Sons Ltd. Published by John Wiley & Sons Ltd.
AB - In this paper, we present a novel approach for the parallel evaluation of procedural shape grammars on the graphics processing unit (GPU). Unlike previous approaches that are either limited in the kind of shapes they allow, the amount of parallelism they can take advantage of, or both, our method supports state of the art procedural modeling including stochasticity and context-sensitivity. To increase parallelism, we explicitly express independence in the grammar, reduce inter-rule dependencies required for context-sensitive evaluation, and introduce intra-rule parallelism. Our rule scheduling scheme avoids unnecessary back and forth between CPU and GPU and reduces round trips to slow global memory by dynamically grouping rules in on-chip shared memory. Our GPU shape grammar implementation is multiple orders of magnitude faster than the standard in CPU-based rule evaluation, while offering equal expressive power. In comparison to the state of the art in GPU shape grammar derivation, our approach is nearly 50 times faster, while adding support for geometric context-sensitivity. © 2014 The Author(s) Computer Graphics Forum © 2014 The Eurographics Association and John Wiley & Sons Ltd. Published by John Wiley & Sons Ltd.
UR - http://hdl.handle.net/10754/563529
UR - http://doi.wiley.com/10.1111/cgf.12312
UR - http://www.scopus.com/inward/record.url?scp=84901852471&partnerID=8YFLogxK
U2 - 10.1111/cgf.12312
DO - 10.1111/cgf.12312
M3 - Article
SN - 0167-7055
VL - 33
SP - 73
EP - 82
JO - Computer Graphics Forum
JF - Computer Graphics Forum
IS - 2
ER -