本案例的目的是理解如何用Metal实现双边模糊效果滤镜,结合图像的空间邻近度和像素值相似度折中处理,同时考虑空域信息和灰度相似性,达到保边去噪的目的;
Demo
实操代码
// 双边模糊滤镜
let filter = C7BilateralBlur.init(radius: 10)
// 方案1:
ImageView.image = try? BoxxIO(element: originImage, filters: [filter, filter2, filter3]).output()
// 方案2:
ImageView.image = originImage.filtering(filter, filter2, filter3)
// 方案3:
ImageView.image = originImage ->> filter ->> filter2 ->> filter3
效果对比图
- 不同参数下效果
| radius: 10 | radius: 2 | radius: 0.5 |
|---|---|---|
|
|
|
实现原理
- 过滤器
这款滤镜采用并行计算编码器设计.compute(kernel: "C7BilateralBlur");
/// 双边模糊
public struct C7BilateralBlur: C7FilterProtocol {
public var radius: Float = 1
public var offect: C7Point2D = C7Point2D.center
public var modifier: Modifier {
return .compute(kernel: "C7BilateralBlur")
}
public var factors: [Float] {
return [radius, offect.x, offect.y]
}
public init(radius: Float = 1) {
self.radius = radius
}
}
- 着色器
将高斯滤波中通过各个点到中心点的空间临近度计算的各个权值进行优化,将其优化为空间临近度(四周8个点)计算的权值和像素值相似度计算的权值的乘积,优化后的权值再与图像作卷积运算;
kernel void C7BilateralBlur(texture2d<half, access::write> outputTexture [[texture(0)]],
texture2d<half, access::sample> inputTexture [[texture(1)]],
constant float *blurRadius [[buffer(0)]],
constant float *stepOffsetX [[buffer(1)]],
constant float *stepOffsetY [[buffer(2)]],
uint2 grid [[thread_position_in_grid]]) {
const int GAUSSIAN_SAMPLES = 9;
const float x = float(grid.x);
const float y = float(grid.y);
const float width = float(inputTexture.get_width());
const float height = float(inputTexture.get_height());
const float2 inCoordinate(x / width, y / height);
int multiplier = 0;
float2 blurStep;
float2 singleStepOffset(float(*stepOffsetX * 10) / width, float(*stepOffsetY * 10) / height);
float2 blurCoordinates[GAUSSIAN_SAMPLES];
for (int i = 0; i < GAUSSIAN_SAMPLES; i++) {
multiplier = (i - ((GAUSSIAN_SAMPLES - 1) / 2));
blurStep = float(multiplier) * singleStepOffset;
blurCoordinates[i] = inCoordinate + blurStep;
}
half4 centralColor;
half gaussianWeightTotal;
half4 sum;
half4 sampleColor;
half distanceFromCentralColor;
half gaussianWeight;
constexpr sampler quadSampler(mag_filter::linear, min_filter::linear);
const float distanceNormalizationFactor = float(abs(1 - *blurRadius));
centralColor = inputTexture.sample(quadSampler, blurCoordinates[4]);
gaussianWeightTotal = 0.18;
sum = centralColor * 0.18;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[0]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.05 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[1]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.09 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[2]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.12 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[3]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.15 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[5]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.15 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[6]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.12 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[7]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.09 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
sampleColor = inputTexture.sample(quadSampler, blurCoordinates[8]);
distanceFromCentralColor = min(distance(centralColor, sampleColor) * distanceNormalizationFactor, 1.0);
gaussianWeight = 0.05 * (1.0 - distanceFromCentralColor);
gaussianWeightTotal += gaussianWeight;
sum += sampleColor * gaussianWeight;
const half4 outColor = sum / gaussianWeightTotal;
outputTexture.write(outColor, grid);
}
Harbeth功能清单
- 支持ios系统和macOS系统
- 支持运算符函数式操作
- 支持多种模式数据源 UIImage, CIImage, CGImage, CMSampleBuffer, CVPixelBuffer.
- 支持快速设计滤镜
- 支持合并多种滤镜效果
- 支持输出源的快速扩展
- 支持相机采集特效
- 支持视频添加滤镜特效
- 支持矩阵卷积
- 支持使用系统 MetalPerformanceShaders.
- 支持兼容 CoreImage.
- 滤镜部分大致分为以下几个模块:
最后
- 慢慢再补充其他相关滤镜,喜欢就给我点个星🌟吧。
-
滤镜Demo地址,目前包含
100+种滤镜,同时也支持CoreImage混合使用。 - 再附上一个开发加速库KJCategoriesDemo地址
- 再附上一个网络基础库RxNetworksDemo地址
- 喜欢的老板们可以点个星🌟,谢谢各位老板!!!
✌️.















网友评论