Class with reference counting wrapping special memory type allocation functions from CUDA.
More...
#include <opencv2/core/cuda.hpp>
|
| HostMem (HostMem::AllocType alloc_type=HostMem::AllocType::PAGE_LOCKED) |
|
| HostMem (const HostMem &m) |
|
| HostMem (int rows, int cols, int type, HostMem::AllocType alloc_type=HostMem::AllocType::PAGE_LOCKED) |
|
| HostMem (Size size, int type, HostMem::AllocType alloc_type=HostMem::AllocType::PAGE_LOCKED) |
|
| HostMem (InputArray arr, HostMem::AllocType alloc_type=HostMem::AllocType::PAGE_LOCKED) |
| creates from host memory with coping data More...
|
|
| ~HostMem () |
|
int | channels () const |
|
HostMem | clone () const |
| returns deep copy of the matrix, i.e. the data is copied More...
|
|
void | create (int rows, int cols, int type) |
| allocates new matrix data unless the matrix already has specified size and type. More...
|
|
void | create (Size size, int type) |
|
GpuMat | createGpuMatHeader () const |
| Maps CPU memory to GPU address space and creates the cuda::GpuMat header without reference counting for it. More...
|
|
Mat | createMatHeader () const |
| returns matrix header with disabled reference counting for HostMem data. More...
|
|
int | depth () const |
|
size_t | elemSize () const |
|
size_t | elemSize1 () const |
|
bool | empty () const |
|
bool | isContinuous () const |
|
HostMem & | operator= (const HostMem &m) |
|
void | release () |
| decrements reference counter and released memory if needed. More...
|
|
HostMem | reshape (int cn, int rows=0) const |
|
Size | size () const |
|
size_t | step1 () const |
|
void | swap (HostMem &b) |
| swaps with other smart pointer More...
|
|
int | type () const |
|
Class with reference counting wrapping special memory type allocation functions from CUDA.
Its interface is also Mat-like but with additional memory type parameters.
- PAGE_LOCKED sets a page locked memory type used commonly for fast and asynchronous uploading/downloading data from/to GPU.
- SHARED specifies a zero copy memory allocation that enables mapping the host memory to GPU address space, if supported.
- WRITE_COMBINED sets the write combined buffer that is not cached by CPU. Such buffers are used to supply GPU with data when GPU only reads it. The advantage is a better CPU cache utilization.
- Note
- Allocation size of such memory types is usually limited. For more details, see CUDA 2.2 Pinned Memory APIs document or CUDA C Programming Guide.
◆ AllocType
Enumerator |
---|
PAGE_LOCKED | |
SHARED | |
WRITE_COMBINED | |
◆ HostMem() [1/5]
cv::cuda::HostMem::HostMem |
( |
HostMem::AllocType |
alloc_type = HostMem::AllocType::PAGE_LOCKED | ) |
|
|
explicit |
Python: |
---|
| cv.cuda.HostMem( | [, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | rows, cols, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | size, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | arr[, alloc_type] | ) -> | <cuda_HostMem object> |
◆ HostMem() [2/5]
cv::cuda::HostMem::HostMem |
( |
const HostMem & |
m | ) |
|
Python: |
---|
| cv.cuda.HostMem( | [, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | rows, cols, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | size, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | arr[, alloc_type] | ) -> | <cuda_HostMem object> |
◆ HostMem() [3/5]
cv::cuda::HostMem::HostMem |
( |
int |
rows, |
|
|
int |
cols, |
|
|
int |
type, |
|
|
HostMem::AllocType |
alloc_type = HostMem::AllocType::PAGE_LOCKED |
|
) |
| |
Python: |
---|
| cv.cuda.HostMem( | [, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | rows, cols, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | size, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | arr[, alloc_type] | ) -> | <cuda_HostMem object> |
◆ HostMem() [4/5]
cv::cuda::HostMem::HostMem |
( |
Size |
size, |
|
|
int |
type, |
|
|
HostMem::AllocType |
alloc_type = HostMem::AllocType::PAGE_LOCKED |
|
) |
| |
Python: |
---|
| cv.cuda.HostMem( | [, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | rows, cols, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | size, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | arr[, alloc_type] | ) -> | <cuda_HostMem object> |
◆ HostMem() [5/5]
Python: |
---|
| cv.cuda.HostMem( | [, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | rows, cols, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | size, type[, alloc_type] | ) -> | <cuda_HostMem object> |
| cv.cuda.HostMem( | arr[, alloc_type] | ) -> | <cuda_HostMem object> |
creates from host memory with coping data
◆ ~HostMem()
cv::cuda::HostMem::~HostMem |
( |
| ) |
|
◆ channels()
int cv::cuda::HostMem::channels |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.channels( | | ) -> | retval |
◆ clone()
HostMem cv::cuda::HostMem::clone |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.clone( | | ) -> | retval |
returns deep copy of the matrix, i.e. the data is copied
◆ create() [1/2]
void cv::cuda::HostMem::create |
( |
int |
rows, |
|
|
int |
cols, |
|
|
int |
type |
|
) |
| |
Python: |
---|
| cv.cuda.HostMem.create( | rows, cols, type | ) -> | None |
allocates new matrix data unless the matrix already has specified size and type.
◆ create() [2/2]
void cv::cuda::HostMem::create |
( |
Size |
size, |
|
|
int |
type |
|
) |
| |
Python: |
---|
| cv.cuda.HostMem.create( | rows, cols, type | ) -> | None |
◆ createGpuMatHeader()
GpuMat cv::cuda::HostMem::createGpuMatHeader |
( |
| ) |
const |
Maps CPU memory to GPU address space and creates the cuda::GpuMat header without reference counting for it.
This can be done only if memory was allocated with the SHARED flag and if it is supported by the hardware. Laptops often share video and CPU memory, so address spaces can be mapped, which eliminates an extra copy.
◆ createMatHeader()
Mat cv::cuda::HostMem::createMatHeader |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.createMatHeader( | | ) -> | retval |
returns matrix header with disabled reference counting for HostMem data.
◆ depth()
int cv::cuda::HostMem::depth |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.depth( | | ) -> | retval |
◆ elemSize()
size_t cv::cuda::HostMem::elemSize |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.elemSize( | | ) -> | retval |
◆ elemSize1()
size_t cv::cuda::HostMem::elemSize1 |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.elemSize1( | | ) -> | retval |
◆ empty()
bool cv::cuda::HostMem::empty |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.empty( | | ) -> | retval |
◆ getAllocator()
◆ isContinuous()
bool cv::cuda::HostMem::isContinuous |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.isContinuous( | | ) -> | retval |
◆ operator=()
◆ release()
void cv::cuda::HostMem::release |
( |
| ) |
|
decrements reference counter and released memory if needed.
◆ reshape()
HostMem cv::cuda::HostMem::reshape |
( |
int |
cn, |
|
|
int |
rows = 0 |
|
) |
| const |
Python: |
---|
| cv.cuda.HostMem.reshape( | cn[, rows] | ) -> | retval |
creates alternative HostMem header for the same data, with different number of channels and/or different number of rows
◆ size()
Size cv::cuda::HostMem::size |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.size( | | ) -> | retval |
◆ step1()
size_t cv::cuda::HostMem::step1 |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.step1( | | ) -> | retval |
◆ swap()
void cv::cuda::HostMem::swap |
( |
HostMem & |
b | ) |
|
Python: |
---|
| cv.cuda.HostMem.swap( | b | ) -> | None |
swaps with other smart pointer
◆ type()
int cv::cuda::HostMem::type |
( |
| ) |
const |
Python: |
---|
| cv.cuda.HostMem.type( | | ) -> | retval |
◆ alloc_type
◆ cols
int cv::cuda::HostMem::cols |
◆ data
uchar* cv::cuda::HostMem::data |
◆ dataend
const uchar* cv::cuda::HostMem::dataend |
◆ datastart
uchar* cv::cuda::HostMem::datastart |
◆ flags
int cv::cuda::HostMem::flags |
◆ refcount
int* cv::cuda::HostMem::refcount |
◆ rows
int cv::cuda::HostMem::rows |
◆ step
size_t cv::cuda::HostMem::step |
The documentation for this class was generated from the following file: