%0 Conference Proceedings %T memCUDA: Map Device Memory to Host Memory on GPGPU Platform %+ School of Computer Science and Technology [Wuhan] %A Jin, Hai %A Li, Bo %A Zhang, Qin %A Ao, Wenbing %< avec comité de lecture %( Lecture Notes in Computer Science %B IFIP International Conference on Network and Parallel Computing (NPC) %C Zhengzhou, China %Y Chen Ding; Zhiyuan Shao; Ran Zheng %I Springer %3 Network and Parallel Computing %V LNCS-6289 %P 299-313 %8 2010-09-13 %D 2010 %R 10.1007/978-3-642-15672-4_26 %K GPU %K CUDA %K memory mapping %K programming model %Z Computer Science [cs]/Digital Libraries [cs.DL]Conference papers %X The Compute Unified Device Architecture (CUDA) programming environment from NVIDIA is a milestone towards making programming many-core GPUs more flexible to programmers. However, there are still many challenges for programmers when using CUDA. One is how to deal with GPU device memory, and data transfer between host memory and GPU device memory explicitly. In this study, source-to-source compiling and runtime library technologies are used to implement an experimental programming system based on CUDA, called memCUDA, which can automatically map GPU device memory to host memory. With some pragma directive language, programmer can directly use host memory in CUDA kernel functions, during which the tedious and error-prone data transfer and device memory management are shielded from programmer. The performance is also improved with some near-optimal technologies. Experiment results show that memCUDA programs can get similar effect with well-optimized CUDA programs with more compact source code. %G English %2 https://inria.hal.science/hal-01058920/document %2 https://inria.hal.science/hal-01058920/file/memCUDA_npc_final.pdf %L hal-01058920 %U https://inria.hal.science/hal-01058920 %~ IFIP-LNCS %~ IFIP %~ IFIP-LNCS-6289 %~ IFIP-NPC %~ IFIP-2010