Resource Type

Journal Article 1136

Conference Information 49

Conference Videos 42

Conference Topics 1

Year

2024 4

2023 124

2022 144

2021 149

2020 119

2019 124

2018 82

2017 80

2016 66

2015 20

2014 13

2013 8

2012 27

2011 11

2010 9

2009 16

2008 22

2007 46

2006 24

2005 28

open ︾

Keywords

Machine learning 42

Deep learning 34

Artificial intelligence 20

neural network 17

Reinforcement learning 14

Neural network 8

Software-defined networking (SDN) 7

cyberspace 5

cyberspace security 5

6G 4

Blockchain 4

Feature selection 4

simulation 4

5G 3

ANN 3

Active learning 3

Additive manufacturing 3

Game theory 3

Network security 3

open ︾

Search scope:

排序: Display mode:

Representation learning via a semi-supervised stacked distance autoencoder for image classification Research Articles

Liang Hou, Xiao-yi Luo, Zi-yang Wang, Jun Liang,jliang@zju.edu.cn

Frontiers of Information Technology & Electronic Engineering 2020, Volume 21, Issue 7,   Pages 963-1118 doi: 10.1631/FITEE.1900116

Abstract: is an important application of deep learning. In a typical classification task, the classification accuracy is strongly related to the features that are extracted via deep learning methods. An is a special type of , often used for dimensionality reduction and feature extraction. The proposed method is based on the traditional , incorporating the “distance” information between samples from different categories. The model is called a semi-supervised distance . Each layer is first pre-trained in an unsupervised manner. In the subsequent supervised training, the optimized parameters are set as the initial values. To obtain more suitable features, we use a stacked model to replace the basic structure with a single hidden layer. A series of experiments are carried out to test the performance of different models on several datasets, including the MNIST dataset, street view house numbers (SVHN) dataset, German traffic sign recognition benchmark (GTSRB), and CIFAR-10 dataset. The proposed semi-supervised distance method is compared with the traditional , sparse , and supervised . Experimental results verify the effectiveness of the proposed model.

Keywords: 自动编码器;图像分类;半监督学习;神经网络    

Deep 3D reconstruction: methods, data, and challenges Review Article

Caixia Liu, Dehui Kong, Shaofan Wang, Zhiyong Wang, Jinghua Li, Baocai Yin,lcxxib@emails.bjut.edu.cn,wangshaofan@bjut.edu.cn

Frontiers of Information Technology & Electronic Engineering 2021, Volume 22, Issue 5,   Pages 615-766 doi: 10.1631/FITEE.2000068

Abstract: Three-dimensional (3D) reconstruction of shapes is an important research topic in the fields of computer vision, computer graphics, pattern recognition, and virtual reality. Existing 3D reconstruction methods usually suffer from two bottlenecks: (1) they involve multiple manually designed states which can lead to cumulative errors, but can hardly learn semantic features of 3D shapes automatically; (2) they depend heavily on the content and quality of images, as well as precisely calibrated cameras. As a result, it is difficult to improve the reconstruction accuracy of those methods. 3D reconstruction methods based on deep learning overcome both of these bottlenecks by automatically learning semantic features of 3D shapes from low-quality images using deep networks. However, while these methods have various architectures, in-depth analysis and comparisons of them are unavailable so far. We present a comprehensive survey of 3D reconstruction methods based on deep learning. First, based on different deep learning model architectures, we divide 3D reconstruction methods based on deep learning into four types, , , , and based methods, and analyze the corresponding methodologies carefully. Second, we investigate four representative databases that are commonly used by the above methods in detail. Third, we give a comprehensive comparison of 3D reconstruction methods based on deep learning, which consists of the results of different methods with respect to the same database, the results of each method with respect to different databases, and the robustness of each method with respect to the number of views. Finally, we discuss future development of 3D reconstruction methods based on deep learning.

Keywords: 深度学习模型;三维重建;循环神经网络;深度自编码器;生成对抗网络;卷积神经网络    

NGAT: attention in breadth and depth exploration for semi-supervised graph representation learning Research Articles

Jianke HU, Yin ZHANG,yinzh@zju.edu.cn

Frontiers of Information Technology & Electronic Engineering 2022, Volume 23, Issue 3,   Pages 409-421 doi: 10.1631/FITEE.2000657

Abstract: Recently, graph neural networks (GNNs) have achieved remarkable performance in representation learning on graph-structured data. However, as the number of network layers increases, GNNs based on the neighborhood aggregation strategy deteriorate due to the problem of oversmoothing, which is the major bottleneck for applying GNNs to real-world graphs. Many efforts have been made to improve the process of feature information aggregation from directly connected nodes, i.e., breadth exploration. However, these models perform the best only in the case of three or fewer layers, and the performance drops rapidly for deep layers. To alleviate oversmoothing, we propose a nested graph network (NGAT), which can work in a semi-supervised manner. In addition to breadth exploration, a -layer NGAT uses a layer-wise aggregation strategy guided by the mechanism to selectively leverage feature information from the -order neighborhood, i.e., depth exploration. Even with a 10-layer or deeper architecture, NGAT can balance the need for preserving the locality (including root node features and the local structure) and aggregating the information from a large neighborhood. In a number of experiments on standard tasks, NGAT outperforms other novel models and achieves state-of-the-art performance.

Keywords: Graph learning     Semi-supervised learning     Node classification     Attention    

Learning to select pseudo labels: a semi-supervised method for named entity recognition Research Articles

Zhen-zhen Li, Da-wei Feng, Dong-sheng Li, Xi-cheng Lu,lizhenzhen14@nudt.edu.cn,davyfeng.c@gmail.com,dsli@nudt.edu.cn,xclu@nudt.edu.cn

Frontiers of Information Technology & Electronic Engineering 2020, Volume 21, Issue 6,   Pages 809-962 doi: 10.1631/FITEE.1800743

Abstract: models have achieved state-of-the-art performance in (NER); the good performance, however, relies heavily on substantial amounts of labeled data. In some specific areas such as medical, financial, and military domains, labeled data is very scarce, while is readily available. Previous studies have used to enrich word representations, but a large amount of entity information in is neglected, which may be beneficial to the NER task. In this study, we propose a for NER tasks, which learns to create high-quality labeled data by applying a pre-trained module to filter out erroneous pseudo labels. Pseudo labels are automatically generated for and used as if they were true labels. Our semi-supervised framework includes three steps: constructing an optimal single neural model for a specific NER task, learning a module that evaluates pseudo labels, and creating new labeled data and improving the NER model iteratively. Experimental results on two English NER tasks and one Chinese clinical NER task demonstrate that our method further improves the performance of the best single neural model. Even when we use only pre-trained static word embeddings and do not rely on any external knowledge, our method achieves comparable performance to those state-of-the-art models on the CoNLL-2003 and OntoNotes 5.0 English NER tasks.

Keywords: 命名实体识别;无标注数据;深度学习;半监督学习方法    

Two-level hierarchical feature learning for image classification Article

Guang-hui SONG,Xiao-gang JIN,Gen-lang CHEN,Yan NIE

Frontiers of Information Technology & Electronic Engineering 2016, Volume 17, Issue 9,   Pages 897-906 doi: 10.1631/FITEE.1500346

Abstract: In some image classification tasks, similarities among different categories are different and the samples are usually misclassified as highly similar categories. To distinguish highly similar categories, more specific features are required so that the classifier can improve the classification performance. In this paper, we propose a novel two-level hierarchical feature learning framework based on the deep convolutional neural network (CNN), which is simple and effective. First, the deep feature extractors of different levels are trained using the transfer learning method that fine-tunes the pre-trained deep CNN model toward the new target dataset. Second, the general feature extracted from all the categories and the specific feature extracted from highly similar categories are fused into a feature vector. Then the final feature representation is fed into a linear classifier. Finally, experiments using the Caltech-256, Oxford Flower-102, and Tasmania Coral Point Count (CPC) datasets demonstrate that the expression ability of the deep features resulting from two-level hierarchical feature learning is powerful. Our proposed method effectively increases the classification accuracy in comparison with flat multiple classification methods.

Keywords: Transfer learning     Feature learning     Deep convolutional neural network     Hierarchical classification     Spectral clustering    

Neuro-heuristic computational intelligence for solving nonlinear pantograph systems Article

Muhammad Asif Zahoor RAJA, Iftikhar AHMAD, Imtiaz KHAN, Muhammed Ibrahem SYAM, Abdul Majid WAZWAZ

Frontiers of Information Technology & Electronic Engineering 2017, Volume 18, Issue 4,   Pages 464-484 doi: 10.1631/FITEE.1500393

Abstract: We present a neuro-heuristic computing platform for finding the solution for initial value problems (IVPs) of nonlinear pantograph systems based on functional differential equations (P-FDEs) of different orders. In this scheme, the strengths of feed-forward artificial neural networks (ANNs), the evolutionary computing technique mainly based on genetic algorithms (GAs), and the interior-point technique (IPT) are exploited. Two types of mathematical models of the systems are constructed with the help of ANNs by defining an unsupervised error with and without exactly satisfying the initial conditions. The design parameters of ANN models are optimized with a hybrid approach GA–IPT, where GA is used as a tool for effective global search, and IPT is incorporated for rapid local convergence. The proposed scheme is tested on three different types of IVPs of P-FDE with orders 1–3. The correctness of the scheme is established by comparison with the existing exact solutions. The accuracy and convergence of the proposed scheme are further validated through a large number of numerical experiments by taking different numbers of neurons in ANN models.

Keywords: Neural networks     Initial value problems (IVPs)     Functional differential equations (FDEs)     Unsupervised learning     Genetic algorithms (GAs)     Interior-point technique (IPT)    

A Rough Fuzzy Neural Classifier

Zeng Huanglin,Wang Xiao

Strategic Study of CAE 2003, Volume 5, Issue 12,   Pages 60-65

Abstract:

In this paper, the concepts of rough sets are used to define equivalence classes encoding input data sets, and eliminate redundant or insignificant attributes in data sets so that to reduce the complexity of system construction. In order to deal with ill-defined or real experimental data, an input object is represented as a fuzzy variable by fuzzy membership function, and the significant factor of the input feature corresponding to output pattern classification is incorporated to constitute a fuzzy inference so that to enhance nonlinear mapping classification. A new kind of rough fuzzy neural classifier and a learning algorithm with LSE are proposed in this paper. A integration of the merits of fuzzy and neural network technologies can not only accommodate overlapping classification and therefore increase the performance of nonlinear mapping classification, but ensure more efficiently to handle real life ambiguous and changing situations and to achieve tractability, robustness, and low-cost solutions.

Keywords: fuzzy sets     rough sets     neural networks     pattern classification    

Pre-training with asynchronous supervised learning for reinforcement learning based autonomous driving Research Articles

Yunpeng Wang, Kunxian Zheng, Daxin Tian, Xuting Duan, Jianshan Zhou,ypwang@buaa.edu.cn,zhengkunxian@buaa.edu.cn,dtian@buaa.edu.cn,duanxuting@buaa.edu.cn

Frontiers of Information Technology & Electronic Engineering 2021, Volume 22, Issue 5,   Pages 615-766 doi: 10.1631/FITEE.1900637

Abstract: Rule-based autonomous driving systems may suffer from increased complexity with large-scale inter-coupled rules, so many researchers are exploring learning-based approaches. (RL) has been applied in designing autonomous driving systems because of its outstanding performance on a wide variety of sequential control problems. However, poor initial performance is a major challenge to the practical implementation of an RL-based autonomous driving system. RL training requires extensive training data before the model achieves reasonable performance, making an RL-based model inapplicable in a real-world setting, particularly when data are expensive. We propose an asynchronous (ASL) method for the RL-based end-to-end autonomous driving model to address the problem of poor initial performance before training this RL-based model in real-world settings. Specifically, prior knowledge is introduced in the ASL pre-training stage by asynchronously executing multiple processes in parallel, on multiple driving demonstration data sets. After pre-training, the model is deployed on a real vehicle to be further trained by RL to adapt to the real environment and continuously break the performance limit. The presented pre-training method is evaluated on the race car simulator, TORCS (The Open Racing Car Simulator), to verify that it can be sufficiently reliable in improving the initial performance and convergence speed of an end-to-end autonomous driving model in the RL training stage. In addition, a real-vehicle verification system is built to verify the feasibility of the proposed pre-training method in a real-vehicle deployment. Simulations results show that using some demonstrations during a supervised pre-training stage allows significant improvements in initial performance and convergence speed in the RL training stage.

Keywords: 自主驾驶;自动驾驶车辆;强化学习;监督学习    

Ensemble enhanced active learning mixture discriminant analysis model and its application for semi-supervised fault classification Research Article

Weijun WANG, Yun WANG, Jun WANG, Xinyun FANG, Yuchen HE

Frontiers of Information Technology & Electronic Engineering 2022, Volume 23, Issue 12,   Pages 1814-1827 doi: 10.1631/FITEE.2200053

Abstract: As an indispensable part of process monitoring, the performance of relies heavily on the sufficiency of process knowledge. However, data labels are always difficult to acquire because of the limited sampling condition or expensive laboratory analysis, which may lead to deterioration of classification performance. To handle this dilemma, a new strategy is performed in which enhanced is employed to evaluate the value of each unlabeled sample with respect to a specific labeled dataset. Unlabeled samples with large values will serve as supplementary information for the training dataset. In addition, we introduce several reasonable indexes and criteria, and thus human labeling interference is greatly reduced. Finally, the effectiveness of the proposed method is evaluated using a numerical example and the Tennessee Eastman process.

Keywords: Semi-supervised     Active learning     Ensemble learning     Mixture discriminant analysis     Fault classification    

Interactive image segmentation with a regression based ensemble learning paradigm Article

Jin ZHANG, Zhao-hui TANG, Wei-hua GUI, Qing CHEN, Jin-ping LIU

Frontiers of Information Technology & Electronic Engineering 2017, Volume 18, Issue 7,   Pages 1002-1020 doi: 10.1631/FITEE.1601401

Abstract: To achieve fine segmentation of complex natural images, people often resort to an interactive segmentation paradigm, since fully automatic methods often fail to obtain a result consistent with the ground truth. However, when the foreground and background share some similar areas in color, the fine segmentation result of conventional interactive methods usually relies on the increase of manual labels. This paper presents a novel interactive image segmentation method via a regression-based ensemble model with semi-supervised learning. The task is formulated as a non-linear problem integrating two complementary spline regressors and strengthening the robustness of each regressor via semi-supervised learning. First, two spline regressors with a complementary nature are constructed based on multivariate adaptive regression splines (MARS) and smooth thin plate spline regression (TPSR). Then, a regressor boosting method based on a clustering hypothesis and semi-supervised learning is proposed to assist the training of MARS and TPSR by using the region segmentation information contained in unlabeled pixels. Next, a support vector regression (SVR) based decision fusion model is adopted to integrate the results of MARS and TPSR. Finally, the GraphCut is introduced and combined with the SVR ensemble results to achieve image segmentation. Extensive experimental results on benchmark datasets of BSDS500 and Pascal VOC have demonstrated the effectiveness of our method, and the com-parison with experiment results has validated that the proposed method is comparable with the state-of-the-art methods for in-teractive natural image segmentation.

Keywords: Interactive image segmentation     Multivariate adaptive regression splines (MARS)     Ensemble learning     Thin-plate spline regression (TPSR)     Semi-supervised learning     Support vector regression (SVR)    

Self-supervised graph learning with target-adaptive masking for session-based recommendation Research Article

Yitong WANG, Fei CAI, Zhiqiang PAN, Chengyu SONG,wangyitong20@nudt.edu.cn,caifei08@nudt.edu.cn,panzhiqiang@nudt.edu.cn,songchengyu@nudt.edu.cn

Frontiers of Information Technology & Electronic Engineering 2023, Volume 24, Issue 1,   Pages 73-87 doi: 10.1631/FITEE.2200137

Abstract: aims to predict the next item based on a user's limited interactions within a short period. Existing approaches use mainly recurrent neural networks (RNNs) or (GNNs) to model the sequential patterns or the transition relationships between items. However, such models either ignore the over-smoothing issue of GNNs, or directly use cross-entropy loss with a softmax layer for model optimization, which easily results in the over-fitting problem. To tackle the above issues, we propose a self-supervised graph learning with (SGL-TM) method. Specifically, we first construct a global graph based on all involved sessions and subsequently capture the self-supervised signals from the global connections between items, which helps supervise the model in generating accurate representations of items in the ongoing session. After that, we calculate the main supervised loss by comparing the ground truth with the predicted scores of items adjusted by our designed module. Finally, we combine the main supervised component with the auxiliary self-supervision module to obtain the final loss for optimizing the model parameters. Extensive experimental results from two benchmark datasets, Gowalla and Diginetica, indicate that SGL-TM can outperform state-of-the-art baselines in terms of Recall@20 and MRR@20, especially in short sessions.

Keywords: Session-based recommendation     Self-supervised learning     Graph neural networks     Target-adaptive masking    

Marine target detection based on Marine-Faster R-CNN for navigation radar plane position indicator images Research Article

Xiaolong CHEN, Xiaoqian MU, Jian GUAN, Ningbo LIU, Wei ZHOU,cxlcxl1209@163.com,guanjian_68@163.com

Frontiers of Information Technology & Electronic Engineering 2022, Volume 23, Issue 4,   Pages 630-643 doi: 10.1631/FITEE.2000611

Abstract: As a classic deep learning target detection algorithm, Faster R-CNN (region convolutional neural network) has been widely used in high-resolution synthetic aperture radar (SAR) and inverse SAR (ISAR) image detection. However, for most common low-resolution radar , it is difficult to achieve good performance. In this paper, taking PPI images as an example, a method based on the Marine-Faster R-CNN algorithm is proposed in the case of complex background (e.‍g., sea clutter) and target characteristics. The method performs feature extraction and target recognition on PPI images generated by radar echoes with the . First, to improve the accuracy of detecting marine targets and reduce the false alarm rate, Faster R-CNN was optimized as the Marine-Faster R-CNN in five respects: new backbone network, anchor size, dense target detection, data sample balance, and scale normalization. Then, JRC (Japan Radio Co., Ltd.) was used to collect echo data under different conditions to build a marine target dataset. Finally, comparisons with the classic Faster R-CNN method and the constant false alarm rate (CFAR) algorithm proved that the proposed method is more accurate and robust, has stronger generalization ability, and can be applied to the detection of marine targets for . Its performance was tested with datasets from different observation conditions (sea states, radar parameters, and different targets).

Keywords: Marine target detection     Navigation radar     Plane position indicator (PPI) images     Convolutional neural network (CNN)     Faster R-CNN (region convolutional neural network) method    

Attention-based efficient robot grasp detection network Research Article

Xiaofei QIN, Wenkai HU, Chen XIAO, Changxiang HE, Songwen PEI, Xuedian ZHANG,xiaofei.qin@usst.edu.cn,obmmd_zxd@163.com

Frontiers of Information Technology & Electronic Engineering 2023, Volume 24, Issue 10,   Pages 1430-1444 doi: 10.1631/FITEE.2200502

Abstract: To balance the inference speed and detection accuracy of a grasp detection algorithm, which are both important for robot grasping tasks, we propose an ; structured pixel-level grasp detection named the attention-based efficient network (AE-GDN). Three spatial attention modules are introduced in the encoder stages to enhance the detailed information, and three channel attention modules are introduced in the stages to extract more semantic information. Several lightweight and efficient DenseBlocks are used to connect the encoder and paths to improve the feature modeling capability of AE-GDN. A high intersection over union (IoU) value between the predicted grasp rectangle and the ground truth does not necessarily mean a high-quality grasp configuration, but might cause a collision. This is because traditional IoU loss calculation methods treat the center part of the predicted rectangle as having the same importance as the area around the grippers. We design a new IoU loss calculation method based on an hourglass box matching mechanism, which will create good correspondence between high IoUs and high-quality grasp configurations. AE-GDN achieves the accuracy of 98.9% and 96.6% on the Cornell and Jacquard datasets, respectively. The inference speed reaches 43.5 frames per second with only about 1.2×10 parameters. The proposed AE-GDN has also been deployed on a practical robotic arm grasping system and performs grasping well. Codes are available at https://github.com/robvincen/robot_gradethttps://github.com/robvincen/robot_gradet.

Keywords: Robot grasp detection     Attention mechanism     Encoder–     decoder     Neural network    

Interactive medical image segmentation with self-adaptive confidence calibration

沈楚云,李文浩,徐琪森,胡斌,金博,蔡海滨,朱凤平,李郁欣,王祥丰

Frontiers of Information Technology & Electronic Engineering 2023, Volume 24, Issue 9,   Pages 1332-1348 doi: 10.1631/FITEE.2200299

Abstract: Interactive medical image segmentation based on human-in-the-loop machine learning is a novel paradigm that draws on human expert knowledge to assist medical image segmentation. However, existing methods often fall into what we call interactive misunderstanding, the essence of which is the dilemma in trading off short- and long-term interaction information. To better use the interaction information at various timescales, we propose an interactive segmentation framework, called interactive MEdical image segmentation with self-adaptive Confidence CAlibration (MECCA), which combines action-based confidence learning and multi-agent reinforcement learning. A novel confidence network is learned by predicting the alignment level of the action with short-term interaction information. A confidence-based reward-shaping mechanism is then proposed to explicitly incorporate confidence in the policy gradient calculation, thus directly correcting the model’s interactive misunderstanding. MECCA also enables user-friendly interactions by reducing the interaction intensity and difficulty via label generation and interaction guidance, respectively. Numerical experiments on different segmentation tasks show that MECCA can significantly improve short- and long-term interaction information utilization efficiency with remarkably fewer labeled samples. The demo video is available at https://bit.ly/mecca-demo-video.

Keywords: Medical image segmentation     Interactive segmentation     Multi-agent reinforcement learning     Confidence learning     Semi-supervised learning    

Estimating Rainfall Intensity Using an Image-Based Deep Learning Model Article

Hang Yin, Feifei Zheng, Huan-Feng Duan, Dragan Savic, Zoran Kapelan

Engineering 2023, Volume 21, Issue 2,   Pages 162-174 doi: 10.1016/j.eng.2021.11.021

Abstract:

Urban flooding is a major issue worldwide, causing huge economic losses and serious threats to public safety. One promising way to mitigate its impacts is to develop a real-time flood risk management system; however, building such a system is often challenging due to the lack of high spatiotemporal rainfall data. While some approaches (i.e., ground rainfall stations or radar and satellite techniques) are available to measure and/or predict rainfall intensity, it is difficult to obtain accurate rainfall data with a desirable spatiotemporal resolution using these methods. This paper proposes an image-based deep learning model to estimate urban rainfall intensity with high spatial and temporal resolution. More specifically, a convolutional neural network (CNN) model called the image-based rainfall CNN (irCNN) model is developed using rainfall images collected from existing dense sensors (i.e., smart phones or transportation cameras) and their corresponding measured rainfall intensity values. The trained irCNN model is subsequently employed to efficiently estimate rainfall intensity based on the sensors' rainfall images. Synthetic rainfall data and real rainfall images are respectively utilized to explore the irCNN's accuracy in theoretically and practically simulating rainfall intensity. The results show that the irCNN model provides rainfall estimates with a mean absolute percentage error ranging between 13.5% and 21.9%, which exceeds the performance of other state-of-the-art modeling techniques in the literature. More importantly, the main feature of the proposed irCNN is its low cost in efficiently acquiring high spatiotemporal urban rainfall data. The irCNN model provides a promising alternative for estimating urban rainfall intensity, which can greatly facilitate the development of urban flood risk management in a real-time manner.

Keywords: Urban flooding     Rainfall images     Deep learning model     Convolutional neural networks (CNNs)     Rainfall intensity    

Title Author Date Type Operation

Representation learning via a semi-supervised stacked distance autoencoder for image classification

Liang Hou, Xiao-yi Luo, Zi-yang Wang, Jun Liang,jliang@zju.edu.cn

Journal Article

Deep 3D reconstruction: methods, data, and challenges

Caixia Liu, Dehui Kong, Shaofan Wang, Zhiyong Wang, Jinghua Li, Baocai Yin,lcxxib@emails.bjut.edu.cn,wangshaofan@bjut.edu.cn

Journal Article

NGAT: attention in breadth and depth exploration for semi-supervised graph representation learning

Jianke HU, Yin ZHANG,yinzh@zju.edu.cn

Journal Article

Learning to select pseudo labels: a semi-supervised method for named entity recognition

Zhen-zhen Li, Da-wei Feng, Dong-sheng Li, Xi-cheng Lu,lizhenzhen14@nudt.edu.cn,davyfeng.c@gmail.com,dsli@nudt.edu.cn,xclu@nudt.edu.cn

Journal Article

Two-level hierarchical feature learning for image classification

Guang-hui SONG,Xiao-gang JIN,Gen-lang CHEN,Yan NIE

Journal Article

Neuro-heuristic computational intelligence for solving nonlinear pantograph systems

Muhammad Asif Zahoor RAJA, Iftikhar AHMAD, Imtiaz KHAN, Muhammed Ibrahem SYAM, Abdul Majid WAZWAZ

Journal Article

A Rough Fuzzy Neural Classifier

Zeng Huanglin,Wang Xiao

Journal Article

Pre-training with asynchronous supervised learning for reinforcement learning based autonomous driving

Yunpeng Wang, Kunxian Zheng, Daxin Tian, Xuting Duan, Jianshan Zhou,ypwang@buaa.edu.cn,zhengkunxian@buaa.edu.cn,dtian@buaa.edu.cn,duanxuting@buaa.edu.cn

Journal Article

Ensemble enhanced active learning mixture discriminant analysis model and its application for semi-supervised fault classification

Weijun WANG, Yun WANG, Jun WANG, Xinyun FANG, Yuchen HE

Journal Article

Interactive image segmentation with a regression based ensemble learning paradigm

Jin ZHANG, Zhao-hui TANG, Wei-hua GUI, Qing CHEN, Jin-ping LIU

Journal Article

Self-supervised graph learning with target-adaptive masking for session-based recommendation

Yitong WANG, Fei CAI, Zhiqiang PAN, Chengyu SONG,wangyitong20@nudt.edu.cn,caifei08@nudt.edu.cn,panzhiqiang@nudt.edu.cn,songchengyu@nudt.edu.cn

Journal Article

Marine target detection based on Marine-Faster R-CNN for navigation radar plane position indicator images

Xiaolong CHEN, Xiaoqian MU, Jian GUAN, Ningbo LIU, Wei ZHOU,cxlcxl1209@163.com,guanjian_68@163.com

Journal Article

Attention-based efficient robot grasp detection network

Xiaofei QIN, Wenkai HU, Chen XIAO, Changxiang HE, Songwen PEI, Xuedian ZHANG,xiaofei.qin@usst.edu.cn,obmmd_zxd@163.com

Journal Article

Interactive medical image segmentation with self-adaptive confidence calibration

沈楚云,李文浩,徐琪森,胡斌,金博,蔡海滨,朱凤平,李郁欣,王祥丰

Journal Article

Estimating Rainfall Intensity Using an Image-Based Deep Learning Model

Hang Yin, Feifei Zheng, Huan-Feng Duan, Dragan Savic, Zoran Kapelan

Journal Article