A Noise-tolerant Differentiable Learning Approach for Single Occurrence Regular Expression with Interleaving

被引:0
|
作者
Ye, Rongzhen [1 ]
Zhuang, Tianqu [1 ]
Wan, Hai [1 ]
Du, Jianfeng [2 ]
Luo, Weilin [1 ]
Liang, Pingjia [1 ]
机构
[1] Sun Yat Sen Univ, Sch Comp Sci & Engn, Guangzhou, Peoples R China
[2] Guangdong Univ Foreign Studies, Guangzhou Key Lab Multilingual Intelligent Proc, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of learning a single occurrence regular expression with interleaving (SOIRE) from a set of text strings possibly with noise. SOIRE fully supports interleaving and covers a large portion of regular expressions used in practice. Learning SOIREs is challenging because it requires heavy computation and text strings usually contain noise in practice. Most of the previous studies only learn restricted SOIREs and are not robust on noisy data. To tackle these issues, we propose a noise-tolerant differentiable learning approach SOIREDL for SOIRE. We design a neural network to simulate SOIRE matching and theoretically prove that certain assignments of the set of parameters learnt by the neural network, called faithful encodings, are one-to-one corresponding to SOIREs for a bounded size. Based on this correspondence, we interpret the target SOIRE from an assignment of the set of parameters of the neural network by exploring the nearest faithful encodings. Experimental results show that SOIREDL outperforms the state-of-the-art approaches, especially on noisy data.
引用
收藏
页码:4809 / 4817
页数:9
相关论文
共 50 条
  • [21] Noise-Tolerant Hybrid Prototypical Learning with Noisy Web Data
    Liang, Chao
    Zhu, Linchao
    Yang, Zongxin
    Chen, Wei
    Yang, Yi
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (10)
  • [22] Noise-tolerant speech recognition: the SNN-TA approach
    Trentin, E
    Matassoni, M
    INFORMATION SCIENCES, 2003, 156 (1-2) : 55 - 69
  • [23] A Noise-Tolerant Approach to Fuzzy-Rough Feature Selection
    Cornelis, Chris
    Jensen, Richard
    2008 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-5, 2008, : 1600 - +
  • [24] Noise-Tolerant Learning for Audio-Visual Action Recognition
    Han, Haochen
    Zheng, Qinghua
    Luo, Minnan
    Miao, Kaiyao
    Tian, Feng
    Chen, Yan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7761 - 7774
  • [25] Noise-tolerant single photon sensitive three-dimensional imager
    Rehain, Patrick
    Sua, Yong Meng
    Zhu, Shenyu
    Dickson, Ivan
    Muthuswamy, Bharathwaj
    Ramanathan, Jeevanandha
    Shahverdi, Amin
    Huang, Yu-Ping
    NATURE COMMUNICATIONS, 2020, 11 (01)
  • [26] Toward Facial Expression Recognition in the Wild via Noise-Tolerant Network
    Gu, Yu
    Yan, Huan
    Zhang, Xiang
    Wang, Yantong
    Ji, Yusheng
    Ren, Fuji
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2033 - 2047
  • [27] Noise-tolerant single-photon imaging with a superconducting nanowire camera
    Kong, Lingdong
    Zhao, Qingyuan
    Zheng, Kai
    Lu, Haiyangbo
    Chen, Shi
    Tao, Xu
    Wang, Hui
    Hao, Hao
    Wan, Chao
    Tu, Xuecou
    Zhang, Labao
    Jia, Xiaoqing
    Kang, Lin
    Chen, Jian
    Wu, Peiheng
    OPTICS LETTERS, 2020, 45 (24) : 6732 - 6735
  • [28] Overhead-free Noise-tolerant Federated Learning: A New Baseline
    Lin, Shiyi
    Zhai, Deming
    Zhang, Feilong
    Jiang, Junjun
    Liu, Xianming
    Ji, Xiangyang
    MACHINE INTELLIGENCE RESEARCH, 2024, 21 (03) : 526 - 537
  • [29] Noise-tolerant single photon sensitive three-dimensional imager
    Patrick Rehain
    Yong Meng Sua
    Shenyu Zhu
    Ivan Dickson
    Bharathwaj Muthuswamy
    Jeevanandha Ramanathan
    Amin Shahverdi
    Yu-Ping Huang
    Nature Communications, 11
  • [30] Noise-Tolerant Radio Frequency Fingerprinting With Data Augmentation and Contrastive Learning
    Ren, Zhanyi
    Ren, Pinyi
    Xu, Dongyang
    Zhang, Tiantian
    2023 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC, 2023,