删除或更新信息,请邮件至freekaoyan#163.com(#换成@)

深度学习技术在学科融合研究中的应用

本站小编 Free考研考试/2022-01-02

<script type="text/javascript" src="https://cdn.bootcss.com/mathjax/2.7.2-beta.0/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script> <script type='text/x-mathjax-config'> MathJax.Hub.Config({ extensions: ["tex2jax.js"], jax: ["input/TeX", "output/HTML-CSS"], tex2jax: {inlineMath: [ ['$','$'], ["\\(","\\)"] ],displayMath: [ ['$$','$$'], ["\\[","\\]"] ],processEscapes: true}, "HTML-CSS": { availableFonts: ["TeX"] }, "HTML-CSS": {linebreaks: {automatic: true}}, SVG: {linebreaks: {automatic: true}} }); </script> 刘晓东,1,*, 倪浩然,21.中国科学院计算机网络信息中心,信息化战略发展与评估中心,北京 100190
2.华威大学,真实系统数学,英国 考文垂,CV4 7HP

Application of Deep Learning Technology in Discipline Integration Research

Liu Xiaodong,1,*, Ni Haoran,21. Center of Informatization Strategy and Evaluation, Computer Network Information Center, Chinese Academy of Sciences, Beijing 100190, China
2. Mathematics for Real-World Systems CDT, University of Warwick, Coventry CV4 7HP, United Kingdom

收稿日期:2020-07-2网络出版日期:2020-10-20

Corresponding authors: Liu Xiaodong (E-mail:liuxiaodong@cnic.cn)
Received:2020-07-2Online:2020-10-20
作者简介 About authors

Liu Xiaodong is an engineer of CNIC. His research interests are data science and artificial intelligence.
In this paper, he is mainly responsible for the data organization, experimental design and experimental result presentations.
E-mail: liuxiaodong@cnic.cn


Ni Haoran is a master student (leading to Ph.D.) at the Mathematics of Systems CDT at Warwick University. His research interests are distributed at Numerical Analysis, Natural Language Processing, Machine Learning Algorithms and Optimal Transports.
In this paper, he is mainly responsible for the experiments and the analysis of experimental results.
E-mail: Haoran.ni@warwick.ac.uk



摘要
【目的】 我们使用深度学习模型对于文章进行多分类,研究论文发表机构的学科融合的科研现状。 【方法】 我们设计了“多类别分类”模型,并应用卷积神经网络对中国科学院产生的8个不同主题的研究论文摘要进行分类。 【结果】 结果表明,科学研究涉及的学科交叉融合变得日趋紧密。 【结论】 多学科的融合交叉促进了科研产出,该研究可进一步用于科研机构的战略规划部署和评价等问题。
关键词: 文本分类;自然语言处理;卷积神经网络;分类算法

Abstract
[Objective] We use deep learning models to multi-classify articles and analyze the disciplinary integration situation of the corresponding institutions. [Methods] In this paper, we design a one-versus-rest classification model and applied convolutional neural networks to categorize paper abstracts of 8 different main subjects produced by Chinese Academy of the Sciences. [Results] The results show that the cross-integration of disciplines involved in scientific research becomes a more frequent practice and the integration of academic fields are promoting the number of publications of scientific research papers. [Conclusions] This research can benefit the strategic planning and deployment for scientific research institutions.
Keywords:text classification;natural language processing;convolutional neural network;classification algorithm


PDF (9621KB)元数据多维度评价相关文章导出EndNote|Ris|Bibtex收藏本文
本文引用格式
刘晓东, 倪浩然. 深度学习技术在学科融合研究中的应用. 数据与计算发展前沿[J], 2020, 2(5): 99-109 doi:10.11871/jfdc.issn.2096-742X.2020.05.010
Liu Xiaodong, Ni Haoran. Application of Deep Learning Technology in Discipline Integration Research. Frontiers of Data and Computing[J], 2020, 2(5): 99-109 doi:10.11871/jfdc.issn.2096-742X.2020.05.010


开放科学标识码(OSID)

1 Introduction/Overview

The number of published papers contributed by the Chinese Academy of Sciences (CAS) is around 70,0001(1 The data in this paper is collected from the public database of Web of Science.)in the year 2018, which rose by 21.8% year-on-year. Due to the increasing number of papers each year and the disciplinary inter-crossing, artificial classification of papers by disciplines becomes harder and harder. There are several difficulties with the artificial classification. First, research paper contains massive information, which makes the workload much heavier than other classification tasks. Secondly, reading papers from specific disciplines requires professional knowledge. However, hiring experts for the classification task is expensive and impractical. Finally, the inter-crossing of disciplines makes it hard to tell even when the papers are evaluated by experts in the related fields. In the meantime, the demand for classifying papers and trend prediction is rapidly increasing. In this background, text classification in natural language processing (NLP) is considered to solve this problem.

As an essential component in NLP applications, text classification has aroused the interest of many researchers. There are many classic application scenarios for it, such as public opinion monitoring (Bing & Lei, 2012)[1], intelligent recommendation system, information filtering and sentiment analysis (Aggarwal & Zhai, 2012)[2].

One of the main problems in text classification is the word representation. Bag-of-Words (BoW) model (Zhang et al., 2010)[3] is the pioneer in this area, where some designed patterns such as uni-grams, bi-grams and n-grams are extracted as features. With the higher order in n-grams and the complicated feature structure (Post & Bergsma, 2013)[4], the model could consider more contextual information and word orders. However, these models failed in capturing the semantics of the sentences and the data sparsity problem remains unsolved. As a distributed representation of words, pre-trained word embedding offers a way to capture meaningful syntactic and semantic regularities (Mikolov, Chen et al., 2013)[5],
and mitigates the data sparsity problem (Bengio et al.,2003)[6]. As one of the state-of-art method of word embedding, word2vec (Rong, 2014)[7] includes two training models: continuous Bag-of-Word (CBoW) (Mikolov, Chen et al., 2013)[5] and skip-grams (Mikolov, Sutskever et al., 2013)[8]. In this paper, we apply the Global Vectors for Word Representation (GloVe) model proposed by Pennington et al. (2014)[9], which outperforms word2vec on the word representation task.

Another problem is the classifier. There are many machine-learning algorithms that can be used as classifiers, such as logistic regression (LR), naive Bayes (NB) and support vector machine (SVM). However, these methods all have the data sparsity problem and do not have a satisfying performance on NLP tasks. With the development of deep neural networks, convolutional neural networks (CNNs) and recurrent neural networks (RNNs) are widely used in NLP tasks as the state-of-art classifiers due to their exceptional performance. In our research, we build a CNN (Kalchbrenner et al., 2014)[10] model and apply it to the paper classification. The input of the model handles sentences of varying length. The layers include several one-dimensional convolutional layers and a k-max pooling layer. The convolutional layers apply one-dimensional filters across each row of features. The max pooling layer is a non-linear sub-sampling function that returns the maximum of a set of values. A convolutional layer followed by a max pooling layer and/or a non-linearity form a basic block to calculate matrix named feature map. In the input layer, we enrich the representation by computing multiple feature maps with disparate filters. Subsequent layers followed by the input layer also have multiple feature maps computed by convolving filters with all the maps. The weights at these layers form an order-4 tensor. The resulting structure is named as Convolutional Neural Network (Lecun et al., 1998)[11].

The networks are trained on 7,000 abstracts of the research papers from 23 different subjects labelled artificially by experts in CAS. After training, the networks generally surpasses 90% in the prediction accuracy on the hand-labelled test, experimented on the well-trained networks in the published research papers contributed by the CAS from 2012 to 2018. By labelling the papers with the classifiers of selected subjects, the trend of interdisciplinary research is shown in the results.

The outline of the paper is as follows. Section 2 reviews the related works including word representation models, neural networks, optimization of hyper-parameters and related applications of text classification. Section 3 defines the classifier, word representation model, the relevant operators and the layers of CNN and the auto-optimization method we applied in the paper. Section 4 discusses the experiments and the results.

2 Related Works

The text classification tasks mainly include three parts: feature engineering, feature selection and classifiers. The feature engineering was commonly based on the BoW model. Rong (2014)[7] proposed the word2vec method, which is widely accepted as the state-of-the-art for word representation. However, it is limited in utilizing statistical information, since there is no global co-occurrence counts. Pennington et al. (2014)[9] introduced the GloVe model, which combines the advantages of latent semantic analysis (LSA) (Deerwrester et al., 1990)[12] and word2vec. This new word-representation method efficiently improved the accuracy of many basic NLP tasks, especially for limited training dataset. There are also other advanced methods recently, such as Bidirectional Encoder Representation from Transformers (BERT1) (Devlin et al., 2018)[13]. Feature selection is applied for filtering the noisy features and improving the performance of the classifiers. The common feature selection methods include removing stopping words, statistic, mutual information and L1 regularization (Ng, 2004)[14]. For classifiers, there are several machine learning algorithms such as logistic regression (LR), naive bayes (NB), and support vector machine (SVM). However, they all have the data sparsity problem.

2.1 Neural Networks in NLP

With the development of deep neural networks, the data sparsity problem has been well solved. Many neural models have been proposed for word embedding and classification. Recursive Neural network (RecursiveNN) (Socher, Huang et al., 2011; Socher, Pennington et al., 2011; Socher et al., 2013)[15,16,17] captures the semantics of a sentence via a tree structure. However, the time complexity of constructing a textual tree is at least O(n2), which is time-consuming for long documents. The Recurrent Neural Network (RNN) (Elman, 1990)[18] is a model, which analyzes the semantics of the text in fixed-sized hidden layers, with a time complexity O(n). However, it is a biased model where later contents are more dominant than earlier contents. To tackle the bias, the Long Short-Term Memory (LSTM) (Liu et al., 2016)[19] model is proposed, which gives a better performance for long documents. The Convolutional Neural Network (CNN) (Kalchbrenner et al., 2014)[10] is another unbiased model introduced to NLP tasks. Although the time complexity of CNN is O(n) too, we found that empirically it outperforms LSTM in both training time and accuracy. Advanced models include Attention Networks (Abreu et al., 2019)[1], Contextual LSTM (C-LSTM) (Ghosh et al., 2016)[20] and Recurrent CNN (RCNN) (Lai et al., 2015)[21]. We will not discuss them since they are too sophisticated and do not contribute noteworthy improvement in accuracy.

The performance of the model also depends on the choice of optimizer and the hyper-parameters. There are many classic optimizers, such as Batch Gradient Descent (BGD), Stochastic Gradient Descent (SGD), Momentum, Adadelta, Adam. We apply the SGD method in our experiments due to its sensitivity to hyper-parameters and model performance. Zhang and Wallace (2015)[22] researches deeply in the sensitivity analysis and gives some practical advises in the selection of hyper-parameters. However, it requires rich experience in NLP engineering, and consumes massive energy and time. Instead of manual selection, we apply auto-optimization algorithms in our research, which offers a modest performance with time and energy saving. The most direct algorithm is Grid Search. However, it costs too much computing resources. Bergstra and Bengio (2012)[6] proposes Random Search which improves the performance of optimization. Bayesian algorithm (Snoek et al., 2012)[23] is more efficient than Random Search, but it cannot deal with continuous variables. To tackle this problem, we apply Tree-structured Parzen Estimator (Bergstra et al., 2011; Bergstra et al., 2013)[24-25]
for optimizing the hyper-parameters. To exploit the parallelism of the operations, the network is trained on GPUs using Pytorch.

3 Methods

3.1 Classifier

For training sample sets {s1, s2, s3, s4}, where Sn, n=1,2,3,4 is the sample set with the category cn, we define the classifier as an one-versus-rest model in our task.

Training Samples

$\left. \begin{array}{ll} +\qquad - \qquad Classifier\ \ prediction\\ {[S_{1}], [S_{2},S_{3}S_{4}]}\ \ \Longrightarrow\ f_{1}\ \ \qquad - \\ {[S_{2}], [S_{1},S_{3}S_{4}]}\ \ \Longrightarrow\ f_{2}\ \ \qquad - \\ {[S_{3}], [S_{1},S_{2}S_{4}]}\ \ \Longrightarrow\ f_{3}\ \ \qquad + \\ {[S_{4}], [S_{1},S_{2}S_{3}]}\ \ \Longrightarrow\ f_{4}\ \ \qquad - \\ \end{array} \right\}Final\ Result\ C_{3}$

In our research, we train separate classifiers for different subjects and label each paper with these classifiers according to the above model. If there is more than one positive prediction for the paper, we believe cross-discipline happens between the labelled subjects.

3.2 Word Representation Learning

We introduce GloVe model in this section. The model is proposed by Pennington et al. (2014)[9]. We denote the matrix of word-word co-occurrence counts as X, where the entries Xij tabulate the number of times word j occurs in the context of word i. The number of times any word appears in the context of word i is denoted as $X_{i}=\sum_{k}X_{ik}$ We denote $W\in\mathbb{R}^{d}$ as word vectors and $\widetilde{W}\in\mathbb{R}^{d}$ as separate context word vectors. The model is given as follows:

$J=\sum^{V}_{ij=1}f(X_{ij})(W^{T}_{i}W_{j}+b_{i}+b_{j}-logX_{ij})^{2}$,
where V is the size of the vocabulary, bi and bj are the bias. We practically choose the function:

$f(x)=\begin{cases}(x/x_{max})^{a}, if\ \ x<x_{max}\\ \qquad1,otherwise\end{cases}$
as the weighting function f(Xij). We fix xmax = 100 and α = 3/4 as they give the best performance for our experiments (Mikolov, Chen et al., 2013)[5].

3.3 Convolutional Neural Network

We build a model using a convolutional architecture that alternates 3 or 4 convolutional layers with pooling layers given by k-max pooling. The resulting architecture is the Convolutional Neural Network.

3.3.1 Convolution

For a vector of weights $m\in\mathbb{R}^{m}$and a vector of inputs $s\in\mathbb{R}^{s}$ as the sentence, the one-dimensional convolutional
operation is to take the dot product of m and each m-gram of s to obtain the vector c:

cj = mTsj - m + 1 : j
There are two types of convolution depending on different requirements. The narrow type requires sm and the resulting sequence $c\in\mathbb{R}^{s-m+1}$ where j is from m to s. In the wide type of convolution, s could be smaller than m. The input si are taken to be zero when i<1 or i>s.In the case, it yields the sequence $c\in\mathbb{R}^{s-m+1}$, where the range of the index j is 1 to s + m - 1. Notice that, the sequence c in the narrow convolution is a sub-sequence of which in the wide convolution. Wide convolution has some advantages such as ensuring that all weights in the filter reach the entire sentence and producing a non-empty c which are independent of the width m and the sentence length s all the time. We apply wide convolution operation in our model. For the sentence matrix $s\in\mathbb{R}^{d\times s}$ which is constructed by the embedding $w_{i}\in\mathbb{R}^{d }$ for each word, we obtain a convolutional layer by convolving a matrix with trained weights $m\in\mathbb{R}^{d\times m}$ with the activations at the layer below. The resulting matrix c at each layer has dimensions d×s + m - 1.

3.3.2 Non-Linear Activation Function

After each convolutional layers, a bias $b\in\mathbb{R}^{d }$ and a non-linear activation function g are applied component-wise to the resulting matrix c of the convolution. Define M to be the matrix of diagonals:

M = [diags(m:,1), ... , digas(m:,m))]
where m are the weights of the d filters of the wide convolution. Thus, we obtain:

yj = g(M[sj, ..., sj+m-1]T + b)
This function makes sure that the output of our network is not just another linear combination of the inputs and gives the possibility to generate more complicated functional relationship for different tasks. There are many different activation functions, such as Sigmoid, tanh (Gulcehre et al., 2016; Glorot & Bengio, 2010)[26,27], ReLU (Nair & Hinton, 2010; Glorot et al., 2010)[28,29] and maxout (Goodfellow et al., 2010)[30]. We apply ReLU function due to its faster convergence rate and generally moderate performance in practice.

3.3.3 Max Pooling

Given a number k and a sequence $p\in\mathbb{R}^{p}$, where pk, k-max pooling is the pooling operation that extracts the k largest values of p with the same order and constructs a sub-sequence pkmax. In our experiments, we set k=1 in the k-max pooling layers.

$z=\mathop{}_{maxy_{i}}^{\mathop{}_{n}^{i=1}}$
where the j-th sequence of the matrix z is the maximum in the j-th elements of yi.

By applying the k-max pooling operators after the convolutional layers, it is available to pool the k most active features in p and to capture the information throughout the entire text. It preserves the order of the features and ignores their positions. It also discerns precisely the number of times the feature is highly activated in p and the change of the feature activations across p. There are other kind of pooling operators such as average pooling (Collobert et al., 2011)[31]. We do not apply them here since only a few words are meaningful for capturing the semantics and further classification in the text. After all the convolutional layers and max pooling layers, a fully connected layer followed by the softmax function is applied for converting the output numbers into probabilities, which predicts the classes of the input sentence.

$p_{i}=\frac{exp(z_{i})}{\sum_{j=1}^{n\sum j}exp}$
For binary classifier, we let n=2 in both Equation 6 and Equation 7.

4 Experiments

4.1 Datasets

Our datasets are collected from the internal database of CAS. The datasets include 330,907 published papers of CAS from 2012 to 2018. We randomly take 7,000 abstracts out of the papers as the training samples labelled them with 23 subjects by experts. In the word representation training, our pre-trained vectors for GloVe are 50-dimensional Wikipedia 2014 + Gigaword 5, which are provided in the GloVe project (Pennington et al., 2014)[9]. After that, we classify the remaining papers by 8 main subjects, which are Astronomy, Atmospheric Science, Electronic Science and Technology, Geology, Mathematics, Nuclear Science and Technology, Physics and Computer Science and Technology, using the well-trained models. We investigate the cross-disciplines between Computer Science and Technology and other 7 selected subjects on timeline.

4.2 Hyper-Parameters and Training

In our experiments, the optimization goal is to minimize the difference between the prediction and the true distribution, which includes an L2 regularization term over the parameters. The set of parameters in the optimization are the word embeddings, the filter and the fully connected weights. The network is trained with mini-batches by back-propagation. Stochastic gradient descent is applied here to optimize the training target. The hyper-parameters comprise the batch size, including the learning rate of optimization, the number of negative samples, the number of layers, the kernel number, and the kernel size of each convolutional layer. To optimize these hyper-parameters, the Tree-structured Parzen Estimator Method is applied here as an auto-optimization algorithm.

4.3 Experiment Settings

We preprocess the datasets as follows. We use Natural Language Toolkit (NLTK) to obtain tokens and stems. Stop words and symbols are also removed by this toolkit. We split each dataset into training dataset, validation dataset and testing dataset by the ratio 6:2:2. We apply F1 measure to evaluate the model performance, Precision, Recall and Accuracy to evaluate the classification results. The hyper-parameter settings of the neural networks depend on the used datasets.

We use the CNN model to classify the abstract information of the paper, and use 3 or 4 layers of convolutions to train the observation effect. The algorithm of the training model chooses to use the stochastic gradient descent algorithm. Because the stochastic gradient descent algorithm is more sensitive to hyper-parameters than other optimization algorithms and can get better results by adjusting the parameters.

4.4 Results and Discussion

4.4.1 Interdisciplinary Design

According to the experimental research in this paper, due to the rapid development of computer science informatization, the development of information technology represented by cloud computing, big data, deep learning, etc. has entered an explosive phase since around 2010, and a large number of cutting-edge technologies of computer science have been used for each scientific research in the field. As shown in the Figure 1, the field of physics ranks first in the number of 48,450 papers published in last 6 years. Using the method proposed in 3.1.2 of this paper, 5,254 papers in the field of physics are used to apply the knowledge in the field of computer science and technology. The crossover rate is as high as 11.1%, ahead of the other seven areas of the dataset. In Figure 2, the integration of various fields and computer science and technology is getting closer and closer, and the correlation between various scientific research work and computer science and technology is getting stronger and stronger.

Fig. 1

新窗口打开|下载原图ZIP|生成PPT
Fig. 1Total number of articles



Fig. 2

新窗口打开|下载原图ZIP|生成PPT
Fig. 2Percentage of crosses



4.4.2 Disciplinary development and interdisciplinary

Secondly, in recent years, in the field of physics, the distributed file systems, machine learning, cloud storage and other technologies have been widely used in the field of physics. The number of papers published in this field has increased from 5,252 in 2012 to 9,204 in 2018, the number of interdisciplinary papers in the field has also increased from 578 to 1,072, both of which have shown a consistent trend of overall curve rise; the same is true in the field of atmospheric science, which grew from 3,808 to 8,598 articles. The cross-disciplinary relationship between this field and computer science and technology has also increased from 410 to 1,011. The growth rate is obvious. The focus of electronic science and technology field is from 4,152 in 2012 to 8,531 in 2018, and it is related to computer science. The interdisciplinary division of technology has also grown from 457 to 1,072. Besides, cross-disciplinary relationships over mathematics, astronomy, and other fields are all growing as shown in Figure 3.

Fig.3

新窗口打开|下载原图ZIP|生成PPT
Fig.3Number of crosses



4.4.3 Research Ability and Subject Cross

In the end, according to our findings, the degree of information infrastructure construction of scientific research institutions shows that the degree of information technology infrastructure construction of scientific research institutions has been steadily increasing year by year, and the scientific research capabilities of large scientific research institutions have been greatly improved. It is consistent with the trend of integration of various research fields with computer science and technology, according to which we believe that there is a positive correlation between the ability of scientific research and the degree of interdisciplinary. The scientific research ability and the degree of interdisciplinary promote and grow mutually.

4.4.4 Research Extension

For the subject classification and fusion problem of this paper, we can apply the results to the evaluation of subject development, talent cultivation, resource optimization, and the establishment of college education study curriculum, or the cultivation of compound talents. The research results in this paper can guide scientific research institutions to carry out strategic planning and deployment better. In addition, our research ideas can also be applied to related expansion issues, such as subject type prediction, attribute judgment, differentiation of low-frequency subjects and confusing subjects. Furthermore, the introduction of attention mechanism would train the subject prediction models to handle the problem of complex case classifications based on the description of each case.

5 Conclusion & Discussion

In this paper, we applied Convolutional Neural Network and Recurrent Neural Network in Research Paper Classification Tasks. We researched the published papers of research institutions over the years, marked the papers and selected eight types of them to train. We used the algorithm model of 1VO+GloVE+CNN to preprocess and train the dataset. The classification and interdisciplinary statistics based on these eight subject models were completed to analyze the number, proportion and regularity of papers with interdisciplinary subjects. Through experiments, this paper believes that the intersection of scientific research work and computer science in various fields is becoming closer, and the degree of interdisciplinary is consistently growing with the rising trend of the number of scientific papers published. In future work, we can apply scientific research results to the development of disciplines, personnel training and resource optimization, which will help scientific research institutions carry out strategic planning and deployment. In the future, we will continue to study different areas on this issue.

参考文献 原文顺序
文献年度倒序
文中引用次数倒序
被引期刊影响因子

Bing L & Lei Z. A Survey of Opinion Mining and Sentiment Analysis
[M] . Boston, MA: Springer US, 2012.

[本文引用: 2]

Aggarwal C & Zhai C, A Survey of Text Classification Algorithms
[M] . Boston, MA: Springer US, 2012.

[本文引用: 1]

Zhang Y, Jin R & Zhou Z, Understanding bag-of-words model: a statistical framework
[J] . International Journal of Machine Learning & Cybernetics, 2010,1(1-4):43-52.

[本文引用: 1]

Post M & Bergsma S. Explicit and implicit syntactic features for text classification
[J] . In Proceedings of the 51st annual meeting of the association for computational linguistics, 2013: volume 2: Short papers, 866-872.

[本文引用: 1]

Mikolov T, Chen K & Corrado G, et al. fficient estimation of word representations in vector space
[J/OL]. arXiv preprint arXiv, 2013,1301. 3781.

[本文引用: 3]

Bengio Y, Ducharme R & Vincent P, et al. A neural probabilistic language model
[J]. Journal of machine learning research , 2003: 1137-1155.

[本文引用: 2]

Rong X. word2vec Parameter Learning Explained
[J/OL]. Computer Science 2014,1411_2738.

[本文引用: 2]

Mikolov T, Sutskever I & Chen K, et al. istributed Representations of Words and Phrases and Their Compositionality [J/OL]
arXiv preprint arXiv, 2013,1310. 4546.

[本文引用: 1]

Pennington J, Socher R & Manning C. D., GloVe: Global Vectors for Word Representation
[C]. Empirical methods in natural language processing, 20141532-1543. http:// www.aclweb.Org /anthology/D14-1162.

URL [本文引用: 4]

Kalchbrenner N., Grefenstette E & Blunsom P., A Convolutional Neural Network for Modelling Sentences
[C]. Association for Computational Linguistics (ACL), 2014.

[本文引用: 2]

Lecun Y, Bottou L & Bengio Y, et al. Gradient-based learning applied to document recognition
[C]. Proceedings of the IEEE (1998), 86(11):2278-2324. doi: 10.1109/5.726791.

[本文引用: 1]

Deerwester S, T. Dumais S & Landauer T, et al. Indexing by latent semantic analysis
[J] . Journal of the American Society for Information Science, 1990,41:391-407. doi: http://www.jfdc.cnic.cn/article/2020/1674-9480/10.1002/(sici)1097-4571(199009)41:660;391::aid-asi162;3.0.co;2-9;391::aid-asi162;3.0.co;2-9.

URL [本文引用: 1]

Devlin J, Chang M, Lee K, et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding [J/OL]
arXiv preprint arXiv, 2018,1810_04805.

[本文引用: 1]

Ng A. Y.. Feature selection, L1 vs. L2 regularization, and rotational invariance
[C] . Proceedings of the twenty-first international conference on Machine learning, 2004.

[本文引用: 1]

Socher R., Huang E & Pennington J. et al. Dynamic Pooling and Unfolding Recursive Autoencoders for Paraphrase Detection
[J] . Advances in neural information processing systems, 2011: 24.

URLPMID:25152607 [本文引用: 1]
A challenging problem in estimating high-dimensional graphical models is to choose the regularization parameter in a data-dependent way. The standard techniques include K-fold cross-validation (K-CV), Akaike information criterion (AIC), and Bayesian information criterion (BIC). Though these methods work well for low-dimensional problems, they are not suitable in high dimensional settings. In this paper, we present StARS: a new stability-based method for choosing the regularization parameter in high dimensional inference for undirected graphs. The method has a clear interpretation: we use the least amount of regularization that simultaneously makes a graph sparse and replicable under random sampling. This interpretation requires essentially no conditions. Under mild conditions, we show that StARS is partially sparsistent in terms of graph estimation: i.e. with high probability, all the true edges will be included in the selected model even when the graph size diverges with the sample size. Empirically, the performance of StARS is compared with the state-of-the-art model selection procedures, including K-CV, AIC, and BIC, on both synthetic data and a real microarray dataset. StARS outperforms all these competing procedures.

Socher R, Pennington J & Huang E, et al. PSemi-supervised recursive autoencoders for predicting sentiment distributions
[C] . Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, 2011.

[本文引用: 1]

Socher R, Perelygin A & Wu J, et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank
[C] . Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, 2013.

[本文引用: 1]

Elman J. Finding structure in time
[C] . Cognitive Science, 1990,14(2):179-211. https://onlinelibrary. wiley. com/doi/abs/10.1207/s15516709cog1402_1doi:10.1207/s15516709cog1402\_1

DOI:10.1207/s15516709cog1402_1URL [本文引用: 1]

Liu P, Qiu X & Huang X, Recurrent Neural Network for Text Classification with Multi-Task Learning
[J/OL]. arXiv preprint arXiv, 2016,1605_05101.

[本文引用: 1]

Ghosh S, Vinyals O, Strope B, et al. Contextual LSTM (CLSTM) models for Large scale NLP tasks
[J/OL]. arXiv preprint arXiv, 2016,1602_06291.

[本文引用: 1]

Lai S, Xu L & Liu K, et al. Recurrent Convolutional Neural Networks for Text Classification
[C] . Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015.

[本文引用: 1]

Zhang Y & Wallace B, A Sensitivity Analysis of (and Practitioners' Guide to) Convolutional Neural Networks for Sentence Classification
[J/OL]. arXiv preprint arXiv, 2015,1510_03820.

[本文引用: 1]

Snoek J, Larochelle H & Adams R. P, Practical Bayesian Optimization of Machine Learning Algorithms
[J/OL]. Advances in Neural Information Processing Systems 25, 2012.

[本文引用: 1]

Bergstra J, Bardenet R, Bengio Y, et al. Algorithms for Hyper-Parameter Optimization
[C] . Advances in Neural Information Processing Systems, 2011.



Bergstra J & Cox D. Making a Science of Model Search: Hyperparameter Optimization in Hundreds of Dimensions for Vision Architectures
[C]. Proceedings of the 30th International Conference on Machine Learning 2013: 115-123.



Gulcehre C, Moczulski M, Denil M, et al. Noisy Activation Functions
[C] . International Conference on Machine Learning, 2016.

[本文引用: 1]

Glorot X & Bengio Y, Understanding the difficulty of training deep feedforward neural networks
[C] . Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, 2010,9, 249-256.

[本文引用: 1]

Nair V & Hinton G, E. Rectified Linear Units Improve Restricted Boltzmann Machines
[C]. Proceedings of the 27th International Conference on Machine Learning, 2010.

[本文引用: 1]

Glorot X, Bordes A & Bengio Y, Deep Sparse Rectifier Neural Networks
[C]. Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, 2011: 15.

[本文引用: 1]

Goodfellow J, Warde-Farley D & Mirza M, et al. Maxout Networks
[C]. Proceedings of the 30 th International Conference on Machine Learning, 2013(3)Vol.28:1319-1327. http://dblp.uni-trier.de/db/conf/icml/icml2013.html.

URL [本文引用: 1]

Collobert R, Weston J & Bottou L, et al. Natural Language Processing (Almost) from Scratch
[J] . Journal of Machine Learning Research, 2011,12(1):2493-2537.

[本文引用: 1]

相关话题/学科 中国科学院 科研 数据 本文