how to use gpu for accelerate the recognition process
jean phil from Canada  [3 posts]
3 years
hi there do someone know a way to use gpu(videocard) for accelerating the efficiency of roborealm.i think there is a way with cuda but befor trying i just ask if someone already do something like that because gpu is really the thing that can accelerate video analysing.
Steven Gentner from United States  [1370 posts] 3 years

Yes there is but it requires rewriting most of the way the modules work. For now, we have focused on utilizing more CPUs until the GPU market becomes more accessible in a smaller form factor. We have been looking at devices like the Gizmo which provide GPU type functionality in a small in-expensive packages but have not yet released anything for those architectures.

Is there a particular module that you are wanting to speed up?

jean phil from Canada  [3 posts] 3 years
multiple object recognition if you already make it face recognition too.my project is a jhonny five robot automated able to interact by voice.
Steven Gentner from United States  [1370 posts] 3 years
The object recognition module does utilize multiple CPU's if you have them. Thus you will notice a speedup if you add more CPUs. We currently do have face detection but not face recognition. That is a module still in the works ... You can, however, use the Haar method in the object recognition module to identify specific faces. This is not ideal since rotation is not handled well by this method unless you train it on multiple angles of your face. Again, that method will use more than 1 CPU if available.


This forum thread has been closed due to inactivity (more than 4 months) or number of replies (more than 50 messages). Please start a New Post and enter a new forum thread with the appropriate title.

 New Post   Forum Index