簡體   English   中英

我的iPhone應用程序占用了太多內存

[英]my iphone app taking up so much memory

我使用UICollectionView制作了圖庫應用,但如下所示的分配性能卻很差。

在此處輸入圖片說明

我找不到哪里不好。 我應該在哪里顯式釋放對象? 請讓我知道。

以下代碼對此表示懷疑。

在collectionView中,

- (UICollectionViewCell *)collectionView:(UICollectionView *)collectionView cellForItemAtIndexPath:(NSIndexPath *)indexPath
...
dispatch_async(all_queue, ^{
    ALAssetRepresentation *representation = [asset defaultRepresentation];
    UIImage *image = [UIImage imageWithCGImage:[representation fullResolutionImage]
                                         scale:[representation scale]
                                   orientation:(UIImageOrientation)[representation orientation]];
    NSString *filename = [representation filename];
    NSLog(@"%@", filename);
    NSLog(@"Loaded Image row : %d", indexPath.row);

    vector<cv::Rect> faces = [ImageUtils findFeature:image minsize:MIN_FACE_SIZE
                                         withCascade:face_cascade];

    Mat imageMat = [ImageUtils cvMatFromUIImage:image];
    for(unsigned int i = 0; i < es.size(); ++i) {
        rectangle(imageMat, cv::Point(es[i].x, es[i].y),
                  cv::Point(es[i].x + es[i].width, es[i].y + es[i].height),
                  cv::Scalar(0,255,255),5);
    }
    dispatch_async(dispatch_get_main_queue(), ^{
        [faceImage setImage:[ImageUtils UIImageFromCVMat:imageMat]];
        [cell setNeedsDisplay];
    });
});
return cell;
}

調用方法

+ (cv::Mat)cvMatFromUIImage:(UIImage *)image
{
CGColorSpaceRef colorSpace = CGImageGetColorSpace(image.CGImage);
CGFloat cols = image.size.width;
CGFloat rows = image.size.height;

cv::Mat cvMat(rows, cols, CV_8UC4); // 8 bits per component, 4 channels (color channels + alpha)

CGContextRef contextRef = CGBitmapContextCreate(cvMat.data,                 // Pointer to  data
                                                cols,                       // Width of bitmap
                                                rows,                       // Height of bitmap
                                                8,                          // Bits per component
                                                cvMat.step[0],              // Bytes per row
                                                colorSpace,                 // Colorspace
                                                kCGImageAlphaNoneSkipLast |
                                                kCGBitmapByteOrderDefault); // Bitmap info flags

CGContextDrawImage(contextRef, CGRectMake(0, 0, cols, rows), image.CGImage);
CGContextRelease(contextRef);
CGColorSpaceRelease(colorSpace);

return cvMat;
}

另一種方法

+ (UIImage *)UIImageFromCVMat:(cv::Mat)cvMat
{
NSData *data = [NSData dataWithBytes:cvMat.data length:cvMat.elemSize()*cvMat.total()];
CGColorSpaceRef colorSpace;

if (cvMat.elemSize() == 1) {
    colorSpace = CGColorSpaceCreateDeviceGray();
} else {
    colorSpace = CGColorSpaceCreateDeviceRGB();
}

CGDataProviderRef provider = CGDataProviderCreateWithCFData((__bridge CFDataRef)data);

// Creating CGImage from cv::Mat
CGImageRef imageRef = CGImageCreate(cvMat.cols,                                 //width
                                    cvMat.rows,                                 //height
                                    8,                                          //bits per component
                                    8 * cvMat.elemSize(),                       //bits per pixel
                                    cvMat.step[0],                            //bytesPerRow
                                    colorSpace,                                 //colorspace
                                    kCGImageAlphaNone|kCGBitmapByteOrderDefault,// bitmap info
                                    provider,                                   //CGDataProviderRef
                                    NULL,                                       //decode
                                    false,                                      //should interpolate
                                    kCGRenderingIntentDefault                   //intent
                                    );


// Getting UIImage from CGImage
UIImage *finalImage = [UIImage imageWithCGImage:imageRef];
CGImageRelease(imageRef);
CGDataProviderRelease(provider);
CGColorSpaceRelease(colorSpace);

return finalImage;
}

另一種方法

+(cv::vector<cv::Rect>)findFeature:(UIImage *)image minsize:(cv::Size)minSize withCascade:(CascadeClassifier)cascade
{
vector<cv::Rect> faces;
Mat frame_gray;
Mat imageMat = [ImageUtils cvMatFromUIImage:image];

cvtColor(imageMat, frame_gray, CV_BGRA2GRAY);
equalizeHist(frame_gray, frame_gray);

cascade.detectMultiScale(frame_gray, faces, 1.1, 2, 0 | CV_HAAR_SCALE_IMAGE, minSize);
frame_gray.release();
imageMat.release();

return faces;
}

這是因為您的UImage分辨率太高。 您必須找到一種減小其尺寸的方法。

創建集合視圖單元格時,請使用dequeueReusableCellWithReuseIdentifier

還要調整正在處理的圖像的尺寸,這肯定會減小尺寸。

暫無
暫無

聲明:本站的技術帖子網頁,遵循CC BY-SA 4.0協議,如果您需要轉載,請注明本站網址或者原文地址。任何問題請咨詢:yoyou2525@163.com.

 
粵ICP備18138465號  © 2020-2024 STACKOOM.COM