DirectCompute Optimizations and Best Practices - Nvidia
DirectCompute Optimizations and Best Practices - Nvidia
DirectCompute Optimizations and Best Practices - Nvidia
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
Multi-GPU<br />
CPU<br />
Inter-GPU communication<br />
occurs via host CPU<br />
PCI-E<br />
…<br />
GPU GPU GPU<br />
Task Parallel<br />
Data Parallel<br />
CS “A” CS “B” CS “Z”<br />
CS “A” CS “A” CS “A”<br />
• Multiple GPUs can be used in a single system for task or data<br />
parallel GPU processing<br />
• Host explicitly manages I/O <strong>and</strong> workload for each GPU<br />
• Choose the best split to minimize inter-GPU communication<br />
(must occur via host memory)