Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Save more on your purchases! discount-offer-chevron-icon
Savings automatically calculated. No voucher code required.
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletter Hub
Free Learning
Arrow right icon
timer SALE ENDS IN
0 Days
:
00 Hours
:
00 Minutes
:
00 Seconds
Arrow up icon
GO TO TOP
Hands-On GPU-Accelerated Computer Vision with OpenCV and CUDA

You're reading from   Hands-On GPU-Accelerated Computer Vision with OpenCV and CUDA Effective techniques for processing complex image data in real time using GPUs

Arrow left icon
Product type Paperback
Published in Sep 2018
Publisher Packt
ISBN-13 9781789348293
Length 380 pages
Edition 1st Edition
Languages
Tools
Arrow right icon
Author (1):
Arrow left icon
 Vaidya Vaidya
Author Profile Icon Vaidya
Vaidya
Arrow right icon
View More author details
Toc

Table of Contents (20) Chapters Close

Title Page
Copyright and Credits
Packt Upsell
Contributors
Preface
1. Introducing CUDA and Getting Started with CUDA FREE CHAPTER 2. Parallel Programming using CUDA C 3. Threads, Synchronization, and Memory 4. Advanced Concepts in CUDA 5. Getting Started with OpenCV with CUDA Support 6. Basic Computer Vision Operations Using OpenCV and CUDA 7. Object Detection and Tracking Using OpenCV and CUDA 8. Introduction to the Jetson TX1 Development Board and Installing OpenCV on Jetson TX1 9. Deploying Computer Vision Applications on Jetson TX1 10. Getting Started with PyCUDA 11. Working with PyCUDA 12. Basic Computer Vision Applications Using PyCUDA 1. Assessments 2. Other Books You May Enjoy Index

Summary


This chapter explained the launch of multiple blocks with each having multiple threads from kernel function. It also showed the method of choosing these two parameters for a large value of threads. It also explained the hierarchical memory architecture that can be used by CUDA programs. The memory nearer to thread execution is fast and as we move away from it memories get slower. When multiple threads want to communicate with each other then CUDA provides the flexibility of using shared memory by which threads from same blocks can communicate with each other. When multiple threads use same memory location then there should be synchronization between this memory access otherwise the final result will not be as expected. We have also seen use of atomic operation to accomplish this synchronization. If some parameters are remaining constant throughout the kernel execution then it can be stored in constant memory for speed up. When CUDA programs exhibit a certain communication pattern...

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at £13.99/month. Cancel anytime
Visually different images