Yiwei Chen 1,2Yi He 1,2,*Hong Ye 1Lina Xing 1,2[ ... ]Guohua Shi 1,2,3
Author Affiliations
Abstract
1 Jiangsu Key Laboratory of Medical Optics, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou 215163, P. R. China
2 School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China Hefei 230026, P. R. China
3 Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences, Shanghai 200031, P. R. China
The prediction of fundus fluorescein angiography (FFA) images from fundus structural images is a cutting-edge research topic in ophthalmological image processing. Prediction comprises estimating FFA from fundus camera imaging, single-phase FFA from scanning laser ophthalmoscopy (SLO), and three-phase FFA also from SLO. Although many deep learning models are available, a single model can only perform one or two of these prediction tasks. To accomplish three prediction tasks using a unified method, we propose a unified deep learning model for predicting FFA images from fundus structure images using a supervised generative adversarial network. The three prediction tasks are processed as follows: data preparation, network training under FFA supervision, and FFA image prediction from fundus structure images on a test set. By comparing the FFA images predicted by our model, pix2pix, and CycleGAN, we demonstrate the remarkable progress achieved by our proposal. The high performance of our model is validated in terms of the peak signal-to-noise ratio, structural similarity index, and mean squared error.
Fundus fluorescein angiography image fundus structure image image translation unified deep learning model generative adversarial networks 
Journal of Innovative Optical Health Sciences
2024, 17(3): 2450003
Author Affiliations
Abstract
School of Astronautics, Harbin Institute of Technology, Harbin, Heilongjiang 150000, P. R. China
Photoacoustic imaging (PAI) is a noninvasive emerging imaging method based on the photoacoustic effect, which provides necessary assistance for medical diagnosis. It has the characteristics of large imaging depth and high contrast. However, limited by the equipment cost and reconstruction time requirements, the existing PAI systems distributed with annular array transducers are difficult to take into account both the image quality and the imaging speed. In this paper, a triple-path feature transform network (TFT-Net) for ring-array photoacoustic tomography is proposed to enhance the imaging quality from limited-view and sparse measurement data. Specifically, the network combines the raw photoacoustic pressure signals and conventional linear reconstruction images as input data, and takes the photoacoustic physical model as a prior information to guide the reconstruction process. In addition, to enhance the ability of extracting signal features, the residual block and squeeze and excitation block are introduced into the TFT-Net. For further efficient reconstruction, the final output of photoacoustic signals uses ‘filter-then-upsample’ operation with a pixel-shuffle multiplexer and a max out module. Experiment results on simulated and in-vivo data demonstrate that the constructed TFT-Net can restore the target boundary clearly, reduce background noise, and realize fast and high-quality photoacoustic image reconstruction of limited view with sparse sampling.
Deep learning feature transformation image reconstruction limited-view measurement photoacoustic tomography 
Journal of Innovative Optical Health Sciences
2024, 17(3): 2350028
Author Affiliations
Abstract
1 Institute of Modern Optics, Nankai University, Tianjin Key Laboratory of Micro-scale Optical Information Science and Technology, Tianjin 300350, China
2 Department of Thyroid and Neck Tumor, Tianjin Medical University Cancer Institute and Hospital National Clinical Research Center for Cancer, Key Laboratory of Cancer Prevention and Therapy, Tianjin 300060, China
Limited by the dynamic range of the detector, saturation artifacts usually occur in optical coherence tomography (OCT) imaging for high scattering media. The available methods are difficult to remove saturation artifacts and restore texture completely in OCT images. We proposed a deep learning-based inpainting method of saturation artifacts in this paper. The generation mechanism of saturation artifacts was analyzed, and experimental and simulated datasets were built based on the mechanism. Enhanced super-resolution generative adversarial networks were trained by the clear–saturated phantom image pairs. The perfect reconstructed results of experimental zebrafish and thyroid OCT images proved its feasibility, strong generalization, and robustness.
Optical coherence tomography saturation artifacts deep learning image inpainting 
Journal of Innovative Optical Health Sciences
2024, 17(3): 2350026
Wenhao Tang 1†Qing Yang 1,2,3Hang Xu 1Yiyu Guo 1[ ... ]Xu Liu 2,3,*
Author Affiliations
Abstract
1 Zhejiang Laboratory, Research Center for Frontier Fundamental Studies, Hangzhou, China
2 Zhejiang University, College of Optical Science and Engineering, State Key Laboratory of Extreme Photonics and Instrumentation, Hangzhou, China
3 ZJU-Hangzhou Global Scientific and Technological Innovation Center, Hangzhou, China
4 Shanghai Jiao Tong University, Chip Hub for Integrated Photonics Xplore (CHIPX), Wuxi, China
With the rapid development of sensor networks, machine vision faces the problem of storing and computing massive data. The human visual system has a very efficient information sense and computation ability, which has enlightening significance for solving the above problems in machine vision. This review aims to comprehensively summarize the latest advances in bio-inspired image sensors that can be used to improve machine-vision processing efficiency. After briefly introducing the research background, the relevant mechanisms of visual information processing in human visual systems are briefly discussed, including layer-by-layer processing, sparse coding, and neural adaptation. Subsequently, the cases and performance of image sensors corresponding to various bio-inspired mechanisms are introduced. Finally, the challenges and perspectives of implementing bio-inspired image sensors for efficient machine vision are discussed.
bio-inspired image sensor machine vision layer-by-layer processing sparse coding neural adaptation 
Advanced Photonics
2024, 6(2): 024001
汪崟 1,*蒋峥 1刘斌 2
作者单位
摘要
1 武汉科技大学 信息科学与工程学院,湖北 武汉 430080
2 武汉科技大学 冶金自动化与检测技术教育部工程研究中心,湖北 武汉 430080
针对传统SIFT匹配算法复杂、特征冗余点多、难以满足实时性等问题,本文提出了一种具有局部自适应阈值的SIFT快速图像匹配算法。首先,所提方法在SIFT算法的基础上,对构建的高斯金字塔进行了优化,通过减少金字塔层数来消除冗余特征点以提高检测效率,并根据图像局部对比度来自适应提取FAST算法中的阈值从而实现高质量的特征点检测,筛选出鲁棒性较强的特征点进行更准确的匹配;其次,采用高斯圆形窗口建立32维降维特征向量,提高算法运行效率;最后,根据匹配特征点对之间的几何一致性对特征点进行提纯,有效减少误匹配。实验结果表明,本文方法在匹配精度和运算效率方面的综合表现均优于SIFT算法及其他对比匹配算法,相比传统的SIFT算法,匹配精度提高了约10%,算法运行时间缩短了约49%。在图像发生尺度、旋转以及光照变化的情况下,正确匹配率在93%以上。
SIFT算法 高斯金字塔 自适应阈值 特征描述符 图像匹配 SIFT algorithm Gaussian pyramid adaptive thresholds feature descriptor image matching 
液晶与显示
2024, 39(2): 228
付惠琛 1,2高军伟 1,2,*车鲁阳 1,2
作者单位
摘要
1 青岛大学 自动化学院,山东 青岛 266071
2 山东省工业控制技术重点实验室,山东 青岛 266071
人体姿态估计和动作识别在安防、医疗和运动等领域有着重要的应用价值。为了解决不同背景及角度下各类运动动作的人体姿态估计和动作识别问题,本文提出了一种改进的YOLOv7-POSE算法,并自行拍摄制作各种拍摄角度的数据集进行训练。此算法以YOLOv7为基础,对原始网络模型添加了分类的功能,在Backbone主干网络中引入CA卷积注意力机制,提升了网络在对人体骨骼关节点和动作的分类的重要特征的识别能力。用HorNet网络结构代替原模型的CBS卷积核,提高了模型的人体关键点检测精度和动作分类的准确度。将Head层的空间金字塔池化结构替换为空洞空间金字塔池化结构,提升了检测精度并且加快了模型收敛。将目标检测框的回归函数由CIOU替换为EIOU,提高了坐标回归的精度。设计了两组对照实验,实验结果证明,改进后的YOLOv7-POSE在验证集上的mAP为95.7%,相比于原始YOLOv7算法提高了4%,各类运动动作识别准确率显著上升,在实际推理中的关键点错检、漏检等情况明显减少,关键点位置估计误差明显降低。
图像处理 关键点检测 姿态估计 注意力机制 空洞空间金字塔池化 image processing key point detection pose estimation convolutional attention mechanism atrous spatial pyramid pooling 
液晶与显示
2024, 39(2): 217
作者单位
摘要
1 辽宁工程技术大学 软件学院,辽宁 葫芦岛 125105
2 汕头职业技术学院 计算机系,广东 汕头 515071
现有的层级式文本生成图像的方法在初始图像生成阶段仅使用上采样进行特征提取,上采样过程本质是卷积运算,卷积运算的局限性会造成全局信息被忽略并且远程语义无法交互。虽然已经有方法在模型中加入自注意力机制,但依然存在图像细节缺失、图像结构性错误等问题。针对上述存在的问题,提出一种基于自监督注意和图像特征融合的生成对抗网络模型SAF-GAN。将基于ContNet的自监督模块加入到初始特征生成阶段,利用注意机制进行图像特征之间的自主映射学习,通过特征的上下文关系引导动态注意矩阵,实现上下文挖掘和自注意学习的高度结合,提高低分辨率图像特征的生成效果,后续通过不同阶段网络的交替训练实现高分辨率图像的细化生成。同时加入了特征融合增强模块,通过将模型上一阶段的低分辨率特征与当前阶段的特征进行融合,生成网络可以充分利用低层特征的高语义信息和高层特征的高分辨率信息,更加保证了不同分辨率特征图的语义一致性,从而实现高分辨率的逼真的图像生成。实验结果表明,相较于基准模型(AttnGAN),SAF-GAN模型在IS和FID指标上均有改善,在CUB数据集上的IS分数提升了0.31,FID指标降低了3.45;在COCO数据集上的IS分数提升了2.68,FID指标降低了5.18。SAF-GAN模型能够有效生成更加真实的图像,证明了该方法的有效性。
计算机视觉 生成对抗网络 文本生成图像 CotNet 图像特征融合 computer vision generative adversarial networks text-to-image cotnet image feature fusion 
液晶与显示
2024, 39(2): 180
作者单位
摘要
宁夏大学 物理与电子电气工程学院,宁夏 银川 750021
针对文本生成图像任务中的文本编码器不能深度挖掘文本信息,导致后续生成的图像存在语义不一致的问题,本文提出了一种改进DMGAN模型的文本生成图像方法。首先使用XLnet的预训练模型对文本进行编码,该模型在大规模语料库的预训练之下能够捕获大量文本的先验知识,实现对上下文信息的深度挖掘;然后在DMGAN模型生成图像的初始阶段和图像细化阶段均加入通道注意力模块,突出重要的特征通道,进一步提升生成图像的语义一致性和空间布局合理性,以及模型的收敛速度和稳定性。实验结果表明,所提出模型在CUB数据集上生成的图像相比原DMGAN模型,IS指标提升了0.47,FID指标降低了2.78,充分说明该模型具有更好的跨模态生成能力。
文本生成图像 XLnet模型 生成对抗网络 通道注意力 text-to-image XLnet model generate adversarial networks attention of channel 
液晶与显示
2024, 39(2): 168
作者单位
摘要
1 辽宁工程技术大学 软件学院,辽宁 葫芦岛 125105
2 湖州师范学院 信息工程学院,浙江 湖州 313000
为充分利用高光谱影像中蕴含的空谱特征,提出了一种半监督空谱局部判别分析的高光谱影像特征提取算法(S4LFDA)。鉴于高光谱数据集具有空间一致性,首先将像元进行空间重构,保存高光谱数据的近邻关系;其次引入光谱信息散度重构像元间的相似度;为了充分利用大量无标签样本提高算法性能,采用模糊C均值聚类算法对样本进行聚类分析得到伪标签;然后通过增加规范化项到局部力导引算法(FDA)的类内散度矩阵和类间散度矩阵中,以此保持无标签样本的聚类结构一致性;最后通过局部FDA算法来保持有标签样本类间散度最大化和类内散度最小化并求解最佳投影向量。S4LFDA算法既保持了数据集在光谱域的可分性,又保持了像元在空间区域内的近邻关系,合理利用有标签样本及无标签样本,提高了算法的分类性能。在Pavia University和Indian Pines数据集上进行实验,总体分类精度达到95.60%和94.38%。与其他维数约简算法相比,该算法有效提高了地物分类性能。
高光谱影像 半监督 空谱 判别分析 特征提取 地物分类 hyperspectral image semi-supervision spatial spectrum discriminant analysis feature extraction feature classification 
液晶与显示
2024, 39(2): 131
刘硕 1,2朱疆 1,2,*陈旭东 1,2王重阳 1,2[ ... ]樊凡 1,2
作者单位
摘要
1 北京信息科技大学仪器科学与光电工程学院,北京 102206
2 北京信息科技大学光电测试技术及仪器教育部重点实验室,北京 102206
光学相干层析成像(OCT)是一种高空间分辨率的光学成像方法,可以对生物组织进行非接触、无标记的二维截面和三维体积成像,能为临床疾病的诊断提供具有重要参考价值的影像信息。在传统的台式OCT系统中,扫描探头被固定在工作台上,探头结构较大,灵活性差,不利于深入狭小腔体内部成像或在床旁检测。本团队设计了一种视频引导的手持式高速OCT系统,其手持探头结构紧凑、体积小巧,便于抓取和深入狭小腔体内部;探头内部集成了相机成像功能,可以实时获得成像区域的视频图像,引导OCT成像。该系统的A线扫描速率可以达到200 kHz。为了克服成像过程中的抖动问题,本团队提出了图像自动配准算法,该算法能显著提高图像质量。采用该系统对离体猪眼角膜和离体猪牙齿进行成像,以验证系统的性能。结果显示该系统能够高速获取高分辨的组织图像。
医用光学 光学相干层析成像 手持探头 图像配准 
中国激光
2024, 51(9): 0907015

关于本站 Cookie 的使用提示

中国光学期刊网使用基于 cookie 的技术来更好地为您提供各项服务,点击此处了解我们的隐私策略。 如您需继续使用本网站,请您授权我们使用本地 cookie 来保存部分信息。
全站搜索
您最值得信赖的光电行业旗舰网络服务平台!