q-state Potts model metastability study using optimized GPU-based Monte Carlo algorithms

January 23rd, 2011


We implemented a GPU based parallel code to perform Monte Carlo simulations of the two dimensional q-state Potts model. The algorithm is based on a checkerboard update scheme and assigns independent random number generators to each thread (one thread per spin). The implementation allows to simulate systems up to ~10^9 spins with an average time per spin flip of 0.147ns on the fastest GPU card tested, representing a speedup up to 155x, compared with an optimized serial code running on a standard CPU. The possibility of performing high speed simulations at large enough system sizes allowed us to provide a positive numerical evidence about the existence of metastability on very large systems based on Binder’s criterion, namely, on the existence or not of specific heat singularities at spinodal temperatures different of the transition one.

(Ezequiel E. Ferrero, Juan Pablo De Francesco, Nicolás Wolovick and Sergio A. Cannas: “q-state Potts model metastability study using optimized GPU-based Monte Carlo algorithms”. [arXiv:1101.0876] [code and additional information])

GPU Implementation of Extended Gaussian Mixture Model for Background Subtraction

January 12th, 2011


Although trivial background subtraction (BGS) algorithms (e.g. frame differencing, running average…) can perform quite fast, they are not robust enough to be used in various computer vision problems. Some complex algorithms usually give better results, but are too slow to be applied to real-time systems. We propose an improved version of the Extended Gaussian mixture model that utilizes the computational power of Graphics Processing Units (GPUs) to achieve real-time performance. Experiments show that our implementation running on a low-end GeForce 9600GT GPU provides at least 10x speedup. The frame rate is greater than 50 frames per second (fps) for most of the tests, even on HD video formats.

(Vu Pham, Phong Vo, Vu Thanh Hung and Le Hoai Bac: “GPU Implementation of Extended Gaussian Mixture Model for Background Subtraction”. IEEE International Conference on Computing and Communication Technologies, Research, Innovation, and Vision for the Future (RIVF), 2010. [DOI] [code and additional information])

Fast GPU-based Collision Detection for Deformable Models

December 27th, 2010


We present a fast GPU-based streaming algorithm to perform collision queries between deformable models. Our approach is based on hierarchical culling and reduces the computation to generating different streams. We present a novel stream registration method to compact the streams and efficiently compute the potentially colliding pairs of primitives. We also use a deferred front tracking method to lower the memory overhead. The overall algorithm has been implemented on different GPUs and we have evaluated its performance on non-rigid and deformable simulations. We highlight our speedups over prior GPU-based and CPU-based algorithms. In practice, our algorithm can perform inter-object and intra-object computations on models composed of hundreds of thousands of triangles in tens of milliseconds.

(Min Tang, Dinesh Manocha, Jiang Lin, Ruofeng Tong, Collision-Streams: “Fast GPU-based Collision Detection for Deformable Models”, in Proceedings of ACM SIGGRAPH Symposium on Interactive 3D Graphics and Games (i3D 2011), San Fransisco, CA, Feb. 18-20, 2011. http://gamma.cs.unc.edu/CSTREAMS)

PhD Thesis: Scientific Computing on Heterogeneous Architectures

December 22nd, 2010


The CPU has traditionally been the computational work horse in scientific computing, but we have seen a tremendous increase in the use of accelerators, such as Graphics Processing Units (GPUs), in the last decade. These architectures are used because they consume less power and offer higher performance than equivalent CPU solutions. They are typically also far less expensive, as more CPUs, and even clusters, are required to match their performance. Even though these accelerators are powerful in terms of floating point operations per second, they are considerably more primitive in terms of capabilities. For example, they cannot even open a file on disk without the use of the CPU. Thus, most applications can benefit from using accelerators to perform heavy computation, whilst running complex tasks on the CPU. This use of different compute resources is often referred to as heterogeneous computing, and we explore the use of heterogeneous architectures for scientific computing in this thesis. Through six papers, we present qualitative and quantitative comparisons of different heterogeneous architectures, the use of GPUs to accelerate linear algebra operations in MATLAB, and efficient shallow water simulation on GPUs. Our results show that the use of heterogeneous architectures can give large performance gains.

(André R. Brodtkorb, “Scientific Computing on Heterogeneous Architectures”, Ph.D. thesis, University of Oslo, Faculty of Mathematics and Natural Sciences, 2010, (PDF))

CfP: First ADBIS workshop on GPUs in Databases (GID 2011)

December 22nd, 2010

GiD LogoThe “GPUs in Databases” workshop is devoted to sharing the knowledge related to applying GPUs in database environments and to discuss possible future development of this application domain. The workshop topics include, but are not limited to:

  • GPU based data compression (lossless/lossy compression and decompression, real time compression and decompression of multimedia)
  • GPUs in databases and data warehouses (join processing, data indexing, data aggregation, bulk query processing, analytical query processing)
  • Data mining using GPUs (classification, frequent itemsets and association rules, frequent subgraphs, sequential patterns, clustering, social networks mining, regression)
  • GPUs in streaming databases (query processing in streaming databases, stream compression/decompression)
  • Applications of GPUs in bioinformatics

The workshop will take place on September 19th, 2011 and is co-located with ADBIS 2011 in Vienna, Austria. Submissions are due April 5th, 2011. All of accepted submissions will be published in CEUR workshop proceedings and the best papers will also be published in Lecture Notes in Computer Science and Foundations of Computing and Decision Sciences.

More detailed information can be found at the workshop website http://gid2011.cs.put.poznan.pl.

CfP: AMD Fusion Developer Summit

December 15th, 2010

From a recent announcement:

Calling all software development innovators in general purpose GPU (GPGPU), data parallel and heterogeneous computing. On June 13-16, 2011 AMD will host the AMD Fusion Developer Summit (AFDS) in Bellevue, Washington. The AFDS conference board has issued a call for presentation proposals, inviting creators of next-generation software to share research and development work through presentations based on the latest technical papers or reports.

AFDS will be a great venue for developers, academics and innovative entrepreneurs to network with others engaged in related work, collectively defining the future course of heterogeneous computing. And delivering a presentation offers you the perfect opportunity to advocate programming paradigms or gain support for industry standards.

The submission deadline is Feb. 4 2011, and the full call is available at http://amd-member.com/newsletters/DevCentral/1012.html.

Call for Participation: ASIM Workshop 2011

December 14th, 2010

The ASIM (Arbeitsgruppe Simulation) and the TUM are jointly organizing the ASIM Workshop 2011 at Technische Universität München (TUM) and the Leibniz Supercomputing Centre, Germany. The workshop theme is “Trends in Computational Science and Engineering: Foundations of Modeling and Simulation” and will take place March 14 to March 16, 2011. The conference program consists of two building blocks: contributed talks and an extensive poster session for new and upcoming Ph.D. students. Poster submissions are cordially invited; registration closes February 12, 2011. More information is available at http://www5.in.tum.de/asim2011.html.

Understanding throughput-oriented architectures

November 24th, 2010


For workloads with abundant parallelism, GPUs deliver higher peak computational throughput than latency-oriented CPUs. Key insights of this article: Throughput-oriented processors tackle problems where parallelism is abundant, yielding design decisions different from more traditional latency oriented processors. Due to their design, programming throughput-oriented processors requires much more emphasis on parallelism and scalability than programming sequential processors. GPUs are the leading exemplars of modern throughput-oriented architecture, providing a ubiquitous commodity platform for exploring throughput-oriented programming.

(Michael Garland and David B. Kirk, “Understanding throughput-oriented architectures”, Commununications of the ACM 53(11), 58-66, Nov. 2010. [DOI])

CfP: 2011 Symposium on Application Accelerators in High Performance Computing (SAAHPC’11)

November 24th, 2010

Papers are solicited for the 2011 Symposium on Application Accelerators in High-Performance Computing. Presentations from technology developers and the academic user community are invited on the following topics:

  • novel accelerator processors, systems, and architectures
  • integration of accelerators with high-performance computing systems
  • programming models for accelerator-based computing
  • languages and compilers for accelerator-based computing
  • run-time environments, profiling and debugging tools for accelerator-based computing
  • scientific and engineering applications that use application accelerators

In addition to the general session, submissions are invited for the following domain-specific topics:

  • Computational chemistry on accelerators (Chair: TBD)
  • Lattice QCD (Chair: Steven Gottlieb, Indiana University, Bloomington)
  • Weather and climate modeling (Chair: John Michalakes, National Renewable Energy Laboratory)
  • Bioinformatics (Chair: TBD)

Submissions are due May 6, 2011, and more information can be found at the symposium website www.saahpc.org.

CfP: High Performance Graphics 2011

November 21st, 2010


We are pleased to announce High-Performance Graphics 2011. High Performance Graphics is the leading international forum for performance-oriented graphics systems research including innovative algorithms, efficient implementations, and hardware architecture. The conference brings together researchers, engineers, and architects to discuss the complex interactions of massively parallel hardware, novel programming models, efficient graphics algorithms, and innovative applications. High Performance Graphics was founded in 2009 to synthesize and expand on two important and well-respected conferences in computer graphics:

  • Graphics Hardware: an annual conference focusing on graphics hardware, architecture, and systems since 1986; and
  • Interactive Ray Tracing: an innovative symposium begun in 2006 focusing on the emerging field of interactive ray tracing and global illumination techniques.

By combining and expanding these two communities, we bring to authors and attendees the best of both fields and a conference covering a broad range of interactive 3D graphics systems and algorithm research.

Conference Info

Sponsored by ACM SIGGRAPH and Eurographics (pending)
The program features three days of paper and industry presentations, with ample time for discussions during breaks, lunches, and the conference banquet.
The conference, which will take place on August 5—7, is co-located with ACM SIGGRAPH 2011 in Vancouver, Canada.
The conference website is located at http://www.highperformancegraphics.org/

Papers Track

We invite original and innovative performance-oriented contributions from all areas of graphics, including hardware architectures, rendering, physics, animation, AI, simulation, and data structures, with topics including (but not limited to): Read the rest of this entry »

Page 20 of 58« First...10...1819202122...304050...Last »