This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
Home Products Support Community News


[Xen-users] xen tesla dom0

To: xen-users@xxxxxxxxxxxxxxxxxxx
Subject: [Xen-users] xen tesla dom0
From: Omar Schiaratura <omar@xxxxxxx>
Date: Thu, 10 Dec 2009 10:50:59 +0100
Delivery-date: Thu, 10 Dec 2009 01:51:53 -0800
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-users-request@lists.xensource.com?subject=help>
List-id: Xen user discussion <xen-users.lists.xensource.com>
List-post: <mailto:xen-users@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=unsubscribe>
Sender: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
User-agent: KMail/1.12.3 (Linux/2.6.31-15-generic; KDE/4.3.3; x86_64; ; )
hi all,
I have a problem executing cuda program in a dom0 kernel.
I compiled the nviodia driver following istruction at this link:
driver seem to work properly and x server starts and work,
but when i run a cuda program seems that no data whas written or read into/from 
gpu memory
i used the following program from cuda sdk:
bandwidtTest, deviceQuery, whith the following results:

Running on......                                                                
      device 0:GeForce GTX 280                                                  
Quick Mode                                                                      
Host to Device Bandwidth for Pageable memory                                    
Transfer Size (Bytes)   Bandwidth(MB/s)                                         
 33554432               3679.3                                                  
Quick Mode                                                                      
Device to Host Bandwidth for Pageable memory                                    
Transfer Size (Bytes)   Bandwidth(MB/s)                                         
 33554432               3662.9                                                  
Quick Mode                                                                      
Device to Device Bandwidth                                                      
Transfer Size (Bytes)   Bandwidth(MB/s)                                         
 33554432               6642.3                                                  

&&&& Test PASSED


Device 0: "GeForce GTX 280"
  Major revision number:                         1
  Minor revision number:                         3
  Total amount of global memory:                 1073414144 bytes
  Number of multiprocessors:                     30              
  Number of cores:                               240             
  Total amount of constant memory:               65536 bytes     
  Total amount of shared memory per block:       16384 bytes     
  Total number of registers available per block: 16384           
  Warp size:                                     32              
  Maximum number of threads per block:           512             
  Maximum sizes of each dimension of a block:    512 x 512 x 64  
  Maximum sizes of each dimension of a grid:     65535 x 65535 x 1
  Maximum memory pitch:                          262144 bytes     
  Texture alignment:                             256 bytes        
  Clock rate:                                    1.30 GHz         
  Concurrent copy and execution:                 Yes              


but when i tried dct8x8 and Montecarlo this is the results:
Using device 0: GeForce GTX 280
CUDA sample DCT/IDCT implementation
Loading test image: barbara.bmp... [512 x 512]... Success
Running Gold 1 (CPU) version... Success
Running Gold 2 (CPU) version... Success
Running CUDA 1 (GPU) version... Success
Running CUDA 2 (GPU) version... Success
Dumping result to barbara_gold1.bmp... Success
Dumping result to barbara_gold2.bmp... Success
Dumping result to barbara_cuda1.bmp... Success
Dumping result to barbara_cuda2.bmp... Success

Processing time (CUDA 1) : 17.022999 ms
Processing time (CUDA 2) : 16.991602 ms
PSNR Original    <---> CPU(Gold 1) : 32.777073
PSNR Original    <---> CPU(Gold 2) : 32.777050
PSNR Original    <---> GPU(CUDA 1) : inf
PSNR Original    <---> GPU(CUDA 2) : inf
PSNR CPU(Gold 1) <---> GPU(CUDA 1) : 32.777073
PSNR CPU(Gold 2) <---> GPU(CUDA 2) : 32.777050

TEST FAILED! (CPU and GPU results differ too much)


Using device 0: GeForce GTX 280
Generating input data...
Allocating memory...
Generating normally distributed samples...
Running GPU Monte Carlo...
Options         : 256
Simulation paths: 262144
Time (ms.)      : 34.305000
GPU options per sec.: 7462.468961
GPU Monte Carlo vs. Black-Scholes statistics
L1 norm        : 1.000000E+00
Average reserve: 0.000000
CPU Monte Carlo vs. Black-Scholes statistics...
L1 norm: 2.970427E-06
Average reserve: 0.000000
CPU vs. GPU Monte Carlo statistics...
L1 norm: 1.000000E+00
Shutting down...

did anyone have experience of that problem?

Xen-users mailing list

<Prev in Thread] Current Thread [Next in Thread>
  • [Xen-users] xen tesla dom0, Omar Schiaratura <=