1. 程式人生 > >常用的Emgu CV程式碼(主要有圖片格式轉換,圖片裁剪,圖片翻轉,圖片旋轉和圖片平移等功能)

常用的Emgu CV程式碼(主要有圖片格式轉換,圖片裁剪,圖片翻轉,圖片旋轉和圖片平移等功能)

轉載自部落格

using System;
using System.Drawing;
using Emgu.CV;
using Emgu.CV.CvEnum;
using Emgu.CV.Structure;

namespace ZNLGIS
{
    public class ImageClass
    {
        //圖片裁剪
        public static Image<Bgr, Byte> Cut(Image<Bgr,Byte> image ,Rectangle rectangle)
        {
            System.Drawing.Size roisize = new
Size(260,380); IntPtr dst = CvInvoke.cvCreateImage(roisize, Emgu.CV.CvEnum.IPL_DEPTH.IPL_DEPTH_8U, 3); CvInvoke.cvSetImageROI(image.Ptr, rectangle); CvInvoke.cvCopy(image.Ptr, dst, IntPtr.Zero); return OpenCVEmguCVDotNet.IplImagePointerToEmgucvImage<Bgr, Byte>(dst); } //圖片裁剪
public static Image<Bgr, Byte> Cut2(Image<Bgr,Byte> image,int oldwidth,int oldheight) { int x = image.Width - oldwidth; int y = image.Height - oldheight; System.Drawing.Size roisize = new System.Drawing.Size(oldwidth, oldheight); //要裁剪的圖片大小
IntPtr dst = CvInvoke.cvCreateImage(roisize, Emgu.CV.CvEnum.IPL_DEPTH.IPL_DEPTH_8U, 3); System.Drawing.Rectangle rect = new System.Drawing.Rectangle(x/2, y/2, oldwidth, oldheight); CvInvoke.cvSetImageROI(image.Ptr, rect); CvInvoke.cvCopy(image.Ptr, dst, IntPtr.Zero); return OpenCVEmguCVDotNet.IplImagePointerToEmgucvImage<Bgr, Byte>(dst); } //圖片翻轉 public static Image<Bgr, Byte> FlipImage(Image<Bgr, Byte> image, bool isHorizontal) { if (isHorizontal) { CvInvoke.cvFlip(image.Ptr, IntPtr.Zero, FLIP.HORIZONTAL); } else { CvInvoke.cvFlip(image.Ptr, IntPtr.Zero, FLIP.VERTICAL); } return image; } //圖片旋轉 public static Image<Bgr, Byte> RotateImage(Image<Bgr, Byte> image_old, double angle, bool clockwise) { IntPtr image_temp; double anglerad = Math.PI * (angle / 180); int newwidth = (int)Math.Abs(image_old.Bitmap.Height * Math.Sin(anglerad)) + (int)Math.Abs(image_old.Bitmap.Width * Math.Cos(anglerad)) + 1; int newheight = (int)Math.Abs(image_old.Bitmap.Height * Math.Cos(anglerad)) + (int)Math.Abs(image_old.Bitmap.Width * Math.Sin(anglerad)) + 1; image_temp = CvInvoke.cvCreateImage(new Size(newwidth, newheight), IPL_DEPTH.IPL_DEPTH_8U, 3); CvInvoke.cvZero(image_temp); int flag = -1; if (clockwise) { flag = 1; } float[] m = new float[6]; int w = image_old.Bitmap.Width; int h = image_old.Bitmap.Height; m[0] = (float)Math.Cos(flag * angle * Math.PI / 180); m[1] = (float)Math.Sin(flag * angle * Math.PI / 180); m[3] = -m[1]; m[4] = m[0]; m[2] = w * 0.5f; m[5] = h * 0.5f; unsafe { void* p; IntPtr ptr; fixed (float* pc = m) { p = (void*)pc; ptr = new IntPtr(p); } IntPtr M = CvInvoke.cvMat(2, 3, MAT_DEPTH.CV_32F, ptr); CvInvoke.cvGetQuadrangleSubPix(image_old.Ptr,image_temp,M); } return OpenCVEmguCVDotNet.IplImagePointerToEmgucvImage<Bgr, Byte>(image_temp); } //圖片平移 public static Image<Bgr, Byte> Py(Image<Bgr, Byte> src,int x,int y) { System.Drawing.Size roisize = new Size(src.Width, src.Height); Image<Bgr, Byte> dst = new Image<Bgr, byte>(src.Width, src.Height, new Bgr(Color.Transparent)); int i, j; int w = src.Width; int h = src.Height; if (x >= 0 && y >= 0) { for (i = 0; i < w - x; i++) { for (j = 0; j < h - y; j++) { CvInvoke.cvSet2D(dst, j + y, i + x, CvInvoke.cvGet2D(src, j, i)); } } } else if (x >= 0 && y < 0) { for (i = 0; i < w - x; i++) { for (j = -y; j < h; j++) { CvInvoke.cvSet2D(dst, j + y, i + x, CvInvoke.cvGet2D(src, j, i)); } } } else if (x < 0 && y >= 0) { for (i = -x; i < w; i++) { for (j = 0; j < h - y; j++) { CvInvoke.cvSet2D(dst, j + y, i + x, CvInvoke.cvGet2D(src, j, i)); } } } else { for (i = -x; i < w; i++) { for (j = -y; j < h; j++) { CvInvoke.cvSet2D(dst, j + y, i + x, CvInvoke.cvGet2D(src, j, i)); } } } return OpenCVEmguCVDotNet.IplImagePointerToEmgucvImage<Bgr, Byte>(dst); } } } using System; using System.Drawing; using System.Drawing.Imaging; using System.Runtime.InteropServices; using Emgu.CV; using Emgu.CV.CvEnum; using Emgu.CV.Structure; namespace ZNLGIS { public class OpenCVEmguCVDotNet { /// <summary> /// 將MIplImage結構轉換到IplImage指標; /// 注意:指標在使用完之後必須用Marshal.FreeHGlobal方法釋放。 /// </summary> /// <param name="mi">MIplImage物件</param> /// <returns>返回IplImage指標</returns> public static IntPtr MIplImageToIplImagePointer(MIplImage mi) { IntPtr ptr = Marshal.AllocHGlobal(mi.nSize); Marshal.StructureToPtr(mi, ptr, false); return ptr; } /// <summary> /// 將IplImage指標轉換成MIplImage結構 /// </summary> /// <param name="ptr">IplImage指標</param> /// <returns>返回MIplImage結構</returns> public static MIplImage IplImagePointerToMIplImage(IntPtr ptr) { return (MIplImage)Marshal.PtrToStructure(ptr, typeof(MIplImage)); } /// <summary> /// 將IplImage指標轉換成Emgucv中的Image物件; /// 注意:這裡需要您自己根據IplImage中的depth和nChannels來決定 /// </summary> /// <typeparam name="TColor">Color type of this image (either Gray, Bgr, Bgra, Hsv, Hls, Lab, Luv, Xyz or Ycc)</typeparam> /// <typeparam name="TDepth">Depth of this image (either Byte, SByte, Single, double, UInt16, Int16 or Int32)</typeparam> /// <param name="ptr">IplImage指標</param> /// <returns>返回Image物件</returns> public static Image<TColor, TDepth> IplImagePointerToEmgucvImage<TColor, TDepth>(IntPtr ptr) where TColor : struct, IColor where TDepth : new() { MIplImage mi = IplImagePointerToMIplImage(ptr); return new Image<TColor, TDepth>(mi.width, mi.height, mi.widthStep, mi.imageData); } /// <summary> /// 將IplImage指標轉換成Emgucv中的IImage介面; /// 1通道對應灰度影象,3通道對應BGR影象,4通道對應BGRA影象。 /// 注意:3通道可能並非BGR影象,而是HLS,HSV等影象 /// </summary> /// <param name="ptr">IplImage指標</param> /// <returns>返回IImage介面</returns> public static IImage IplImagePointToEmgucvIImage(IntPtr ptr) { MIplImage mi = IplImagePointerToMIplImage(ptr); Type tColor; Type tDepth; string unsupportedDepth = "不支援的畫素位深度IPL_DEPTH"; string unsupportedChannels = "不支援的通道數(僅支援1,2,4通道)"; switch (mi.nChannels) { case 1: tColor = typeof(Gray); switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: tDepth = typeof(Byte); return new Image<Gray, Byte>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16U: tDepth = typeof(UInt16); return new Image<Gray, UInt16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16S: tDepth = typeof(Int16); return new Image<Gray, Int16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32S: tDepth = typeof(Int32); return new Image<Gray, Int32>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32F: tDepth = typeof(Single); return new Image<Gray, Single>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_64F: tDepth = typeof(Double); return new Image<Gray, Double>(mi.width, mi.height, mi.widthStep, mi.imageData); default: throw new NotImplementedException(unsupportedDepth); } case 3: tColor = typeof(Bgr); switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: tDepth = typeof(Byte); return new Image<Bgr, Byte>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16U: tDepth = typeof(UInt16); return new Image<Bgr, UInt16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16S: tDepth = typeof(Int16); return new Image<Bgr, Int16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32S: tDepth = typeof(Int32); return new Image<Bgr, Int32>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32F: tDepth = typeof(Single); return new Image<Bgr, Single>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_64F: tDepth = typeof(Double); return new Image<Bgr, Double>(mi.width, mi.height, mi.widthStep, mi.imageData); default: throw new NotImplementedException(unsupportedDepth); } case 4: tColor = typeof(Bgra); switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: tDepth = typeof(Byte); return new Image<Bgra, Byte>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16U: tDepth = typeof(UInt16); return new Image<Bgra, UInt16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_16S: tDepth = typeof(Int16); return new Image<Bgra, Int16>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32S: tDepth = typeof(Int32); return new Image<Bgra, Int32>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_32F: tDepth = typeof(Single); return new Image<Bgra, Single>(mi.width, mi.height, mi.widthStep, mi.imageData); case IPL_DEPTH.IPL_DEPTH_64F: tDepth = typeof(Double); return new Image<Bgra, Double>(mi.width, mi.height, mi.widthStep, mi.imageData); default: throw new NotImplementedException(unsupportedDepth); } default: throw new NotImplementedException(unsupportedChannels); } } /// <summary> /// 將Emgucv中的Image物件轉換成IplImage指標; /// </summary> /// <typeparam name="TColor">Color type of this image (either Gray, Bgr, Bgra, Hsv, Hls, Lab, Luv, Xyz or Ycc)</typeparam> /// <typeparam name="TDepth">Depth of this image (either Byte, SByte, Single, double, UInt16, Int16 or Int32)</typeparam> /// <param name="image">Image物件</param> /// <returns>返回IplImage指標</returns> public static IntPtr EmgucvImageToIplImagePointer<TColor, TDepth>(Image<TColor, TDepth> image) where TColor : struct, IColor where TDepth : new() { return image.Ptr; } /// <summary> /// 將IplImage指標轉換成點陣圖物件; /// 對於不支援的畫素格式,可以先使用cvCvtColor函式轉換成支援的影象指標 /// </summary> /// <param name="ptr">IplImage指標</param> /// <returns>返回點陣圖物件</returns> public static Bitmap IplImagePointerToBitmap(IntPtr ptr) { MIplImage mi = IplImagePointerToMIplImage(ptr); PixelFormat pixelFormat; //畫素格式 string unsupportedDepth = "不支援的畫素位深度IPL_DEPTH"; string unsupportedChannels = "不支援的通道數(僅支援1,2,4通道)"; switch (mi.nChannels) { case 1: switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: pixelFormat = PixelFormat.Format8bppIndexed; break; case IPL_DEPTH.IPL_DEPTH_16U: pixelFormat = PixelFormat.Format16bppGrayScale; break; default: throw new NotImplementedException(unsupportedDepth); } break; case 3: switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: pixelFormat = PixelFormat.Format24bppRgb; break; case IPL_DEPTH.IPL_DEPTH_16U: pixelFormat = PixelFormat.Format48bppRgb; break; default: throw new NotImplementedException(unsupportedDepth); } break; case 4: switch (mi.depth) { case IPL_DEPTH.IPL_DEPTH_8U: pixelFormat = PixelFormat.Format32bppArgb; break; case IPL_DEPTH.IPL_DEPTH_16U: pixelFormat = PixelFormat.Format64bppArgb; break; default: throw new NotImplementedException(unsupportedDepth); } break; default: throw new NotImplementedException(unsupportedChannels); } Bitmap bitmap = new Bitmap(mi.width, mi.height, mi.widthStep, pixelFormat, mi.imageData); //對於灰度影象,還要修改調色盤 if (pixelFormat == PixelFormat.Format8bppIndexed) SetColorPaletteOfGrayscaleBitmap(bitmap); return bitmap; } /// <summary> /// 將點陣圖轉換成IplImage指標 /// </summary> /// <param name="bitmap">點陣圖物件</param> /// <returns>返回IplImage指標</returns> public static IntPtr BitmapToIplImagePointer(Bitmap bitmap) { IImage iimage = null; switch (bitmap.PixelFormat) { case PixelFormat.Format8bppIndexed: iimage = new Image<Gray, Byte>(bitmap); break; case PixelFormat.Format16bppGrayScale: iimage = new Image<Gray, UInt16>(bitmap); break; case PixelFormat.Format24bppRgb: iimage = new Image<Bgr, Byte>(bitmap); break; case PixelFormat.Format32bppArgb: iimage = new Image<Bgra, Byte>(bitmap); break; case PixelFormat.Format48bppRgb: iimage = new Image<Bgr, UInt16>(bitmap); break; case PixelFormat.Format64bppArgb: iimage = new Image<Bgra, UInt16>(bitmap); break; default: Image<Bgra, Byte> tmp1 = new Image<Bgra, Byte>(bitmap.Size); Byte[, ,] data = tmp1.Data; for (int i = 0; i < bitmap.Width; i++) { for (int j = 0; j < bitmap.Height; j++) { Color color = bitmap.GetPixel(i, j); data[j, i, 0] = color.B; data[j, i, 1] = color.G; data[j, i, 2] = color.R; data[j, i, 3] = color.A; } } iimage = tmp1; break; } return iimage.Ptr; } /// <summary> /// 設定256級灰度點陣圖的調色盤 /// </summary> /// <param name="bitmap"></param> public static void SetColorPaletteOfGrayscaleBitmap(Bitmap bitmap) { PixelFormat pixelFormat = bitmap.PixelFormat; if (pixelFormat == PixelFormat.Format8bppIndexed) { ColorPalette palette = bitmap.Palette; for (int i = 0; i < palette.Entries.Length; i++) palette.Entries[i] = Color.FromArgb(255, i, i, i); bitmap.Palette = palette; } } } }