Hi Everyone,
I'm working on combining two different pieces of code for a ORB Slam project and am trying to figure out how to re-format the 'descriptors'.
The FROM data is this,
std::vector descriptors;
(I'm then adding to this)
The TO format is this,
_descriptors.create(descriptors.size(), 32, CV_8U);
How can I copy this for best performance?
Cheers
Fred
↧
How to copy std::vector to pre-formatted Mat!?
↧
createVideoReader > 4096 px
Hello,
I am currently developing a program for the distribution of video (or others) on giant LED screens. Some video exceeds 4096px and I can't open them via the "cv::cudacodec::createVideoReader" function
Do you have an idea to solve my problem?
I thank you in advance
ps : Here is the error :
Assertion failed (videoFormat.width >= decodeCaps.nMinWidth && videoFormat.height >= decodeCaps.nMinHeight && videoFormat.width <= decodeCaps.nMaxWidth && videoFormat.height <= decodeCaps.nMaxHeight) in cv::cudacodec::detail::VideoDecoder::create
↧
↧
cv2.imread() gives None, but the path/cwd is correct
Hello,
I had this problem already and it was solved without any clear reason. I had my Jupyter Notebook tab open and didn't change anything and then it worked. I already read articles and other threads on this, but couldn't find a solution for this; just look at my screenshot:

On a PyImageSearch blog post I read that this can happen with bad compiling, such as JPEGs. But this doesn't make sense: It already worked a few hours ago and I could display the img with plt.imshow(). (Not with cv2.imshow(), that leads to my kernel crashing.)
↧
How to convert color Image to PPM image to use in TensorRT
Hi, im working on a program that uses TensorRT API to run an inference with ssd_mobilnet model. My problem is that TensorRT takes ppm format image and must of the forums I have search uses "imwrite" to save the image to ppm but i need the image in memory like a cv::Mat matrix so i dont have to keep writing into disk every converted frame.
Thanks for any response.
↧
Unhandled exception (msvcr120.dll) using imwrite()
I'm developing a software that has a few threads writing 10 to 50 bitmap files per second and sometimes I'm getting this exception.Each thread writes its bitmap files into different folders.
I tried cloning the cv::Mat just before imwrite(), or getting a lock before calling imwrite() but those didn't help.
Each threads uses this code to write bitmap files.
cv::Mat source_image; // UINT8, 1x128x130
// do some processing to the source_image;
cv::Mat cloned_image = source_image.clone();
if ((cloned_image.rows > 0) && (cloned_image.cols >0)) {
cv::String file_name = ""; // somewhere
cv::imwrite(file_name, cloned_image); // THIS IS WHERE I GET THE EXCEPTION
}
I also checked the mini dump file (*.dmp) and the exception code was 0xC0000409.
Here is the stack info from the mini dump file.
msvcr120.dll!_invoke_watson(const wchar_t * pszExpression, const wchar_t * pszFunction, const wchar_t * pszFile, unsigned int nLine, unsigned __int64 pReserved) Line 132 C++
msvcr120.dll!_invalid_parameter(const wchar_t * pszExpression, const wchar_t * pszFunction, const wchar_t * pszFile, unsigned int nLine, unsigned __int64 pReserved) Line 85 C++
msvcr120.dll!_invalid_parameter_noinfo() Line 97 C++
msvcr120.dll!_write(int fh, const void * buf, unsigned int cnt) Line 66 C
msvcr120.dll!_fwrite_nolock(const void * buffer, unsigned __int64 size, unsigned __int64 num, _iobuf * stream) Line 166 C
msvcr120.dll!fwrite(const void * buffer, unsigned __int64 size, unsigned __int64 count, _iobuf * stream) Line 83 C
opencv_world300.dll!00007ffc0e265373() unknown
opencv_world300.dll!00007ffc0e264745() unknown
opencv_world300.dll!00007ffc0e256899() unknown
opencv_world300.dll!00007ffc0e2532bb() unknown
opencv_world300.dll!00007ffc0e252f36() unknown
cv::imwrite(file_name, cloned_image);
Thank you.
↧
↧
How to link files to use GpuMat in Visual Studio?
I want to use CUDA/GPU in OpenCV in Visual Studio. For example, `cuda::GpuMat`. After I successfully build OpenCV with the extra modules with CUDA enabled, am I supposed to go to `Property Pages` in VS, then add additional files/directories in `C/C++->General` and `Linker->General` and `Input`?
↧
vignetting correction
Hello,
I want to do 16bit image vignetting correction and for this, i used below link
https://github.com/GUOYI1/Vignetting_corrector/tree/master/C%2B%2B/Vig_Correction
This is in type IPLImage, but now IPLImage is outdated that's why I converted into Mat type .
This logic is worked correctly for 8bit image but when I converted for 16bit image(changes like unsigned char to unsigned short) its not working properly.
Please give me any suggetions.
↧
AdobeRGB to RGB without change in color or making it washed out
This is more like a query.I am processing some images using opencv and the images are in AdobeRBG space , opencv is able to convert it to RGB .But thats deteriorating the image color.the later one look dull. Is there a way I can handle this case.
↧
How to get accurate transformation between two 2D point cloud?
I am using OpenCV ICP algorithm to find Transformation matrix(pose) between one scan 2D data with respect to reference 2D data (CAD DATA) of object.
Reference data: 
Scan data: 
**Approach to filter out points:**
Here for better matching, as shown in below image one imaginary straight points are created(black) in reference data and only points used which are right side of the red line (left sides of points are variable).
filtered reference data: 
Also, in scan data as shown in below image points taken which are right side of the line and imaginary points created. Additionally to match points at the starting (because of convex) in scan data some straight points are created at top (black dots in image) in straight lines(using highest points from that region) and some additional points were removed.
filtered scan data: 
But, OpenCV ICP is not giving accurate result for translation and rotation every time. is there any way to get better transformation matrix or pose for datasets of 2D point cloud? what can be possible approach with or without ICP?
↧
↧
Hi, I would like to ask if OAK-1 is compatible with Arduino?
Does it have an I2C interface or something like that?
↧
Inspect video from ZIP in Python
I have a number of zip files with videos and other files. I'm interested in extracting the duration of each video, in order to calculate the total duration per zip file. I found this article on Kaggle: [https://www.kaggle.com/humananalog/examine-mp4-files-with-python-only](https://www.kaggle.com/humananalog/examine-mp4-files-with-python-only). It should work, but it gets stuck in some MP4s.
My main obstacle to using openCV so far is that it doesn't allow me to read from a buffer. It's locked to either file name or device number. I found an answer that seems to mention some read_buffer function, but it is available under C++ and I'm doing Python.
This is what I got so far:
from zipfile import ZipFile as ZF
file = 'path/to/file.zip'
z = ZF(file,'r')
count = 0
total = 0
for info in z.infolist():
if info.filename.endswith('.mp4'):
# pp(list(map(lambda a: {a: info.__getattribute__(a)}, info.__dir__())))
count += 1
print(f'{count}. {info.filename}')
vid = z.read(info.filename)
print(len(vid))
total += len(vid)
print(f'Read {total} bytes in {count} files.')
Is there any way to read that `vid` buffer into `cv2` and get its duration in seconds in any way that doesn't involve saving a file to disk and reading it back through `cv2.VideoCapture`? I'm using openCV 4.2.0 under Python 3.7.6 and I could update, if needed.
Many thanks!
↧
"undefined reference to ..." OpenCV_3.4.10 on ubuntu 18.04
Hello,
I have errors when I'm trying to compile this program :
#include
#include
#include
#include
int main()
{
std::cout << "Hello world" << std::endl;
cv::Mat image = cv::imread("/media/D280-BD55/circleGrid_6x4.jpg", CV_LOAD_IMAGE_COLOR);
}
I have these errors :
g++ main.cpp -lopencv_core -lopencv_highgui -lopencv_imgproc -o main
/tmp/ccyRyEy9.o: In function `main':
main.cpp:(.text+0x6c): undefined reference to `cv::imread(cv::String const&, int)'
/tmp/ccyRyEy9.o: In function `cv::String::String(char const*)':
main.cpp:(.text._ZN2cv6StringC2EPKc[_ZN2cv6StringC5EPKc]+0x4c): undefined reference to `cv::String::allocate(unsigned long)'
/tmp/ccyRyEy9.o: In function `cv::String::~String()':
main.cpp:(.text._ZN2cv6StringD2Ev[_ZN2cv6StringD5Ev]+0x10): undefined reference to `cv::String::deallocate()'
collect2: error: ld returned 1 exit status
The odd thing is that if I comment the cv::imread() instruction, it works fine.
I tryed a lot of things I found on the web but nothing worked for me
↧
show video and texts on single window
Hi,
I would like to show a video stream and sensors data to same window. But video should be displayed in Left half of the LCD and Sensors data should be displayed on right half of the LCD.
I am using 800x480 resolution LCD to display. I could show video and text on same video using opencv. But I don't have idea about above thing.
Please be needful.
I am newbie to opencv, so might be couldn't explain you clearly. Please find below code till now.
import cv2
import sys
import numpy as np
import logging as log
import datetime as dt
from time import sleep
import imutils
log.basicConfig(filename='webcam.log',level=log.INFO)
b,g,r,a = 0,255,0,0
video_capture = cv2.VideoCapture(0)
anterior = 0
while True:
if not video_capture.isOpened():
print('Unable to load camera.')
sleep(5)
pass
# Capture frame-by-frame
ret, frame = video_capture.read()
frame = imutils.resize(frame, width=800,height=480)
text = "Hello"
cv2.putText(frame,text,(100,100),cv2.FONT_HERSHEY_SIMPLEX, 1.5, (b,g,r),1,cv2.LINE_AA)
if cv2.waitKey(1) & 0xFF == ord('q'):
break
# Display the resulting frame
cv2.imshow('Video', frame)
# When everything is done, release the capture
video_capture.release()
cv2.destroyAllWindows()
Thanks in advance.
Jitendra
↧
↧
How to use OpenCV 4 in Java on Raspberry PI (armv6 and armv7 Raspian)?
Please note this is a cross-post also present on [stackoverflow](https://stackoverflow.com/questions/62505966/how-to-use-opencv-4-in-java-on-raspberry-pi-armv6-and-armv7-raspian), but I would also like the expertise of this wonderful expert community.
This question is very similar to [this one][1] however it's on Linux ARM ( Raspberry Pi).
I've compiled OpenCV 4.4.0 from source along with the Java bindings and tried something like this hack (that worked on Windows):
import org.opencv.core.*;
public class CVTest {
public static void main(String[] args) {
System.out.println("setup");
// loading the typical way fails :(
// System.loadLibrary(Core.NATIVE_LIBRARY_NAME);
System.load("/home/pi/eclipse/CVTest/lib/libopencv_core.so");
////System.load("/home/pi/opencv/build/lib/libopencv_core.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_imgproc.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_imgcodecs.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_img_hash.so");
////System.load("/home/pi/eclipse/CVTest/lib/opencv_core.so");//videoio_ffmpeg440_64.dll
//System.load("/home/pi/eclipse/CVTest/lib/libopencv_videoio.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_photo.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_xphoto.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_flann.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_features2d.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_calib3d.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_phase_unwrapping.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_structured_light.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_xfeatures2d.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_video.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_ximgproc.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_aruco.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_bgsegm.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_bioinspired.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_objdetect.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_face.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_dnn.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_tracking.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_plot.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_ml.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_ml.so");
System.load("/home/pi/eclipse/CVTest/lib/libopencv_text.so");
// crashes here!
System.load("/home/pi/eclipse/CVTest/lib/libopencv_java440.so");
Mat m = Mat.eye(new Size(3,3), CvType.CV_8UC1);
System.out.println("done");
}
}
However, depending on the Raspberry Pi, I get different crashes at the same line, loading `libopencv_java440` (after the other dependant libraries have loaded):
On a Raspberry Pi 3B running Raspian stretch I get errors like this one:
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGBUS (0x7) at pc=0x6360f644, pid=9730, tid=0x64eba470
#
# JRE version: Java(TM) SE Runtime Environment (8.0_202-b08) (build 1.8.0_202-b08)
# Java VM: Java HotSpot(TM) Client VM (25.202-b08 mixed mode linux-arm )
# Problematic frame:
# C [libopencv_core.so+0x258644] cv::Ptr, std::allocator>, std::allocator, std::allocator>>>>::~Ptr()+0x38
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# If you would like to submit a bug report, please visit:
# http://bugreport.java.com/bugreport/crash.jsp
#
--------------- T H R E A D ---------------
Current thread (0x76162400): VMThread [stack: 0x64e3b000,0x64ebb000] [id=9733]
siginfo: si_signo: 7 (SIGBUS), si_code: 1 (BUS_ADRALN), si_addr: 0x7c1f010e
Registers:
r0 = 0x636e6270
r1 = 0x00000000
r2 = 0x000011f8
r3 = 0x7c1f010e
r4 = 0x7c1f00f2
r5 = 0x636e6270
r6 = 0x76ee84ac
r7 = 0x00000001
r8 = 0x6470c600
r9 = 0x000003ec
r10 = 0x000003ec
fp = 0x64eb9c0c
r12 = 0x76fa4ce4
sp = 0x64eb9be8
lr = 0x76ddadd4
pc = 0x6360f644
cpsr = 0x20000010
Top of Stack: (sp=0x64eb9be8)
0x64eb9be8: 76ee8000 00000000 76ee84ac 76ddadd4
0x64eb9bf8: 76129d7c 76ddae34 00000000 76635d7c
0x64eb9c08: 64eb9c2c 768ac3dc 76162270 63736954
0x64eb9c18: 76940000 76129988 76129990 76129d7c
0x64eb9c28: 64eb9ca4 768aaa54 7696a050 76942d14
0x64eb9c38: 64eb9ca4 767f8084 00000000 00000000
0x64eb9c48: 8c365cd8 00000000 76909901 76966044
0x64eb9c58: 76163208 00000000 00000000 7696a050
Instructions: (pc=0x6360f644)
0x6360f624: 0a00000d e59f20c4 e7933002 e3530000
0x6360f634: 0a00000b e284301c ee070fba e1932f9f
0x6360f644: e2421001 e1830f91 e3500000 1afffffa
0x6360f654: e3520001 ee070fba 0a00000f e1a00005
Register to memory mapping:
r0 = 0x636e6270
On a Raspberry Pi ZeroW also running Raspian Stretch I a beefy log: [example][2].
Any tips on getting OpenCV Java bindings to work on arvm6/armv7 CPUs ?
I've ran the tests as well.
As you can see in [opencv_cpp_tests.txt][3] mosts tests run, excluding ones where asset loading is required (Must've botched running the asset part).
I did try running the Java tests as well, however I'm missing something obvious because the `java.library.path` argument I pass to the lib folder containing opencv's shared libraries doesn't seem to work. You can view the output in [opencv_java_tests.txt](https://lifesine.eu/so/opencv_java_tests.txt)
I've also tried the old-school 2.4 [Introduction to Java Development OpenCV Tutorial][4] specifying the correct java class paths and library path, but got greeted by a segfault :/
java -Djava.library.path=/home/pi/opencv/build/lib -classpath /home/pi/opencv/build/bin/opencv-440.jar:/home/pi/opencv/build/build/jar/SimpleSample.jar SimpleSample -verbose
Segmentation fault
**Update**
Using static or shared libs still segfaults :(
To ease testing for anyone out there with a Raspberry PI 3 B+ I've uploaded the following:
- [HelloCV.tar.gz][5]: contains Hello.java (which simply tried to load the library, create a `Mat` and print it), opencv-440.jar, libopencv_java440.so (static build) and compile.sh and run.sh which should call `javac`/`java` with the right arguments, locally referencing the java and c++ libraries.
- [opencv440_static_rpi3b.tar.gz](http://lifesine.eu/so/opencv440_static_rpi3b.tar.gz): static library build
- [opencv440_shared_rpi3b.tar.gz](http://lifesine.eu/so/opencv440_shared_rpi3b.tar.gz): shared library build
Additionally I will try recompiling without libatomic nor `NEON` and `VPF3` CPU optimisations and will post updates
Your advice/tips are very much appreciated ! Thank you
[1]: https://stackoverflow.com/questions/62503875/how-to-use-opencv-4-in-java-on-windows
[2]: https://pastebin.com/RaVH0SsD
[3]: https://lifesine.eu/so/opencv_cpp_tests.txt
[4]: https://docs.opencv.org/2.4/doc/tutorials/introduction/desktop_java/java_dev_intro.html
↧
Time of Flight camera with ROS driver.
Hi,
I have a Time of Flight camera that has a ROS driver. Would this be enough to get the video feed and be able to use the OpenCV tools for tasks such as this one? https://www.youtube.com/watch?v=We6CQHhhOFo
Thanks in advance!
↧
Why is elemSize 6 and depth 2 of rgb Mat after CV_BayerBG2BGR ?
When using OpenCV RAW to RGB conversion (CV_BayerBG2BGR), the image is displayed correctly but the image basic types are incorrect (elemSize & depth).
Even writing the converted Mat to file and then loading it (rgb_path in code below), there is a discrepancy between the loaded rgb image and the converted rgb image, although both display fine.
This causes an issue downstream where I convert from Mat to uint8_t* as the buffer size is larger in the converted rgb image.
Is this an issue with the conversion itself or my understanding of the conversion / OpenCV basic data types? I am using OpenCV 341.
int main() {
Mat img = imread(rgb_path);
ifstream ifd(raw_path, ios::binary | ios::ate);
int size = ifd.tellg();
ifd.seekg(0, ios::beg);
vector buffer;
buffer.resize(size);
ifd.read(buffer.data(), size);
Mat rgb_image;
Mat raw_image(600, 800, CV_16UC1, buffer.data());
cvtColor(raw_image, rgb_image, CV_BayerBG2BGR);
cout << "elemSize() orig: " << img.elemSize() << endl;
cout << "elemSize() conv: " << rgb_image.elemSize() << endl;
cout << "channels() conv: " << rgb_image.channels() << endl;
cout << "channels() orig: " << img.channels() << endl;
cout << "depth() conv: " << rgb_image.depth() << endl;
cout << "depth() orig: " << img.depth() << endl;
return 0;
}
Output:
elemSize() orig: 3
elemSize() conv: 6
channels() conv: 3
channels() orig: 3
depth() conv: 2
depth() orig: 0
↧
Opencv.js build does not contain cv.matfromarray
I followed [this tutorial][1] and built opencv.js. When I try to use the generated file I get the error that cv.MatFromArray is undefined.
You can see the contents of build_js.py I use in the gist below (I use it unedited in the build process); the commands that cmake takes lets it include or exclude certain modules.
https://gist.github.com/tomatopuree/be5b59868d1bcb10b317b7f839fd2b08
What is confusing is that the core libraries, which contain cv.MatFromArray are included in the build.
Therefore I should not be getting this error message! Please help.
The command I use to build opencv.js (although probably not relevant to this problem since the build succeeds):
python ./platforms/js/build_js.py build_wasm --build_wasm --emscripten_dir=/opencv_builds/emsdk/upstream/emscripten
[1]: https://docs.opencv.org/master/d4/da1/tutorial_js_setup.html
↧
↧
Image data of dtype complex128 cannot be converted to float. how to remove this error?
.
from numpy import mean,cov,cumsum,dot,linalg,size,flipud
import numpy as np
import matplotlib.pyplot as plt
import numpy
#from ipymol.compat import Image
from PIL import Image
import cv2
def princomp(A,numpc=0):
# computing eigenvalues and eigenvectors of covariance matrix
M = (A-mean(A.T,axis=1)).T # subtract the mean (along columns)
[latent,coeff] = linalg.eig(cov(M))
p = size(coeff,axis=1)
idx = np.argsort(latent) # sorting the eigenvalues
idx = idx[::-1] # in ascending order
# sorting eigenvectors according to the sorted eigenvalues
coeff = coeff[:,idx]
latent = latent[idx] # sorting eigenvalues
if numpc < p and numpc >= 0:
coeff = coeff[:,range(numpc)] # cutting some PCs if needed
score = dot(coeff.T,M) # projection of the data in the new space
return coeff,score,latent
from pylab import imread,subplot,imshow,title,gray,figure,show,NullLocator
A = cv2.imread('C:\\Users\\dell\\OneDrive\\Desktop\\Image1.png') # load an image
A = mean(A,2) # to get a 2-D array
full_pc = size(A,axis=1) # numbers of all the principal components
i = 1
dist = []
for numpc in range(0,full_pc+10,10): # 0 10 20 ... full_pc
coeff, score, latent = princomp(A,numpc)
Ar = dot(coeff,score).T+mean(A,axis=0) # image reconstruction
# difference in Frobenius norm
dist.append(linalg.norm(A-Ar,'fro'))
# showing the pics reconstructed with less than 50 PCs
if numpc <= 50:
ax = subplot(2,3,i,frame_on=False)
ax.xaxis.set_major_locator(NullLocator()) # remove ticks
ax.yaxis.set_major_locator(NullLocator())
i += 1
#imshow(flipud(Ar))
#plt.imshow(np.flipud(Ar))
title('PCs # '+str(numpc))
gray()
#figure()
#imshow(flipud(A))
gray()
show()
#plt.imshow(np.flipud(A))
title('numpc FULL')
plt.show()
from pylab import plot,axis
figure()
perc = cumsum(latent)/sum(latent)
dist = dist/max(dist)
plot(range(len(perc)),perc,'b',range(0,full_pc+10,10),dist,'r')
axis([0,full_pc,0,1.1])
show()
↧
Why cv:: encode gives similar size of output buffer for bgr8 and mono8 images?
Hello
I notice **4.1 MB/s** @50fps 1280x720 bgr8.
.....and **3.7 MB/s** @50fps 1280x720 mono8.
Above test performed with same camera looking at same scene in same ambient light conditions.
I was expecting around 3 times less bandwidth consumption for grayscale.
jai@jai:~$ opencv_version
4.2.0
jai@jai:~$ opencv_version -v | grep jpeg
JPEG: /usr/lib/x86_64-linux-gnu/libjpeg.so (ver 80)
jai@jai:~$ ldconfig -p | grep jpeg
libmjpegutils-2.1.so.0 (libc6,x86-64) => /lib/x86_64-linux-gnu/libmjpegutils-2.1.so.0
libjpegxr.so.0 (libc6,x86-64) => /lib/x86_64-linux-gnu/libjpegxr.so.0
libjpeg.so.8 (libc6,x86-64) => /lib/x86_64-linux-gnu/libjpeg.so.8
libjpeg.so (libc6,x86-64) => /lib/x86_64-linux-gnu/libjpeg.so
libgdcmjpeg16.so.3.0 (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg16.so.3.0
libgdcmjpeg16.so (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg16.so
libgdcmjpeg12.so.3.0 (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg12.so.3.0
libgdcmjpeg12.so (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg12.so
libgdcmjpeg8.so.3.0 (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg8.so.3.0
libgdcmjpeg8.so (libc6,x86-64) => /lib/x86_64-linux-gnu/libgdcmjpeg8.so
jai@jai:~$
FYI: I am using ros::compressed_image_transport plugin.
I checked on opencv4 and opencv3.4.9.
↧
how to find example media files
HI I'm working on some example python script:
https://docs.opencv.org/3.0-beta/doc/py_tutorials/py_video/py_meanshift/py_meanshift.html#meanshift
and it calls for a certain slow.flv
I can't find it any where on my PC where I installed python and opencv, so surely there is a depository of all the example media like photos and videos on the official opencv website.
Anyone know?
thanks
↧