<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "http://dtd.nlm.nih.gov/publishing/2.0/journalpublishing.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" article-type="research-article" dtd-version="2.0">
  <front>
    <journal-meta>
      <journal-id journal-id-type="publisher-id">JMIR</journal-id>
      <journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id>
      <journal-title>Journal of Medical Internet Research</journal-title>
      <issn pub-type="epub">1438-8871</issn>
      <publisher>
        <publisher-name>JMIR Publications</publisher-name>
        <publisher-loc>Toronto, Canada</publisher-loc>
      </publisher>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="publisher-id">v21i5e12957</article-id>
      <article-id pub-id-type="pmid">31127715</article-id>
      <article-id pub-id-type="doi">10.2196/12957</article-id>
      <article-categories>
        <subj-group subj-group-type="heading">
          <subject>Original Paper</subject>
        </subj-group>
        <subj-group subj-group-type="article-type">
          <subject>Original Paper</subject>
        </subj-group>
      </article-categories>
      <title-group>
        <article-title>The Deep Learning–Based Recommender System “Pubmender” for Choosing a Biomedical Publication Venue: Development and Validation Study</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="editor">
          <name>
            <surname>Parra-Calderón</surname>
            <given-names>Carlos Luis</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Lin</surname>
            <given-names>Chin</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Zhang</surname>
            <given-names>Liangliang</given-names>
          </name>
        </contrib>
        <contrib contrib-type="reviewer">
          <name>
            <surname>Brenas</surname>
            <given-names>Jon</given-names>
          </name>
        </contrib>
      </contrib-group>
      <contrib-group>
        <contrib contrib-type="author" id="contrib1">
          <name name-style="western">
            <surname>Feng</surname>
            <given-names>Xiaoyue</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0003-3954-1333</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib2">
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>Hao</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-4606-8727</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib3">
          <name name-style="western">
            <surname>Ren</surname>
            <given-names>Yijie</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-4867-2338</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib4">
          <name name-style="western">
            <surname>Shang</surname>
            <given-names>Penghui</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0001-8908-7410</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib5">
          <name name-style="western">
            <surname>Zhu</surname>
            <given-names>Yi</given-names>
          </name>
          <degrees>BSc</degrees>
          <xref rid="aff2" ref-type="aff">2</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0001-7115-2544</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib6">
          <name name-style="western">
            <surname>Liang</surname>
            <given-names>Yanchun</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-1147-3968</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib7" equal-contrib="yes">
          <name name-style="western">
            <surname>Guan</surname>
            <given-names>Renchu</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <xref rid="aff3" ref-type="aff">3</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-7162-7826</ext-link>
        </contrib>
        <contrib contrib-type="author" id="contrib8" corresp="yes" equal-contrib="yes">
          <name name-style="western">
            <surname>Xu</surname>
            <given-names>Dong</given-names>
          </name>
          <degrees>PhD</degrees>
          <xref rid="aff1" ref-type="aff">1</xref>
          <address>
            <institution>Key Laboratory of Symbolic Computation and Knowledge Engineering of the Ministry of Education</institution>
            <institution>College of Computer Science and Technology</institution>
            <institution>Jilin University</institution>
            <addr-line>2699 Qinjin Street</addr-line>
            <addr-line>Changchun, 130012</addr-line>
            <country>China</country>
            <phone>86 13944088266</phone>
            <email>xudong@missouri.edu</email>
          </address>
          <xref rid="aff2" ref-type="aff">2</xref>
          <xref rid="aff4" ref-type="aff">4</xref>
          <ext-link ext-link-type="orcid">http://orcid.org/0000-0002-4809-0514</ext-link>
        </contrib>
      </contrib-group>
      <aff id="aff1">
      <label>1</label>
      <institution>Key Laboratory of Symbolic Computation and Knowledge Engineering of the Ministry of Education</institution>
      <institution>College of Computer Science and Technology</institution>  
      <institution>Jilin University</institution>  
      <addr-line>Changchun</addr-line>
      <country>China</country></aff>
      <aff id="aff2">
      <label>2</label>
      <institution>Department of Electric Engineering and Computer Science</institution>
      <institution>University of Missouri</institution>  
      <addr-line>Columbia, MO</addr-line>
      <country>United States</country></aff>
      <aff id="aff3">
      <label>3</label>
      <institution>Zhuhai Sub Laboratory, Key Laboratory of Symbolic Computation and Knowledge Engineering of the Ministry of Education</institution>
      <institution>Zhuhai College</institution>  
      <institution>Jilin University</institution>  
      <addr-line>Zhuhai</addr-line>
      <country>China</country></aff>
      <aff id="aff4">
      <label>4</label>
      <institution>Christopher S Bond Life Sciences Center</institution>
      <institution>University of Missouri</institution>  
      <addr-line>Columbia, MO</addr-line>
      <country>United States</country></aff>
      <author-notes>
        <corresp>Corresponding Author: Dong Xu 
        <email>xudong@missouri.edu</email></corresp>
      </author-notes>
      <pub-date pub-type="collection"><month>05</month><year>2019</year></pub-date>
      <pub-date pub-type="epub">
        <day>24</day>
        <month>05</month>
        <year>2019</year>
      </pub-date>
      <volume>21</volume>
      <issue>5</issue>
      <elocation-id>e12957</elocation-id>
      <!--history from ojs - api-xml-->
      <history>
        <date date-type="received">
          <day>28</day>
          <month>11</month>
          <year>2018</year>
        </date>
        <date date-type="rev-request">
          <day>31</day>
          <month>3</month>
          <year>2019</year>
        </date>
        <date date-type="accepted">
          <day>19</day>
          <month>4</month>
          <year>2019</year>
        </date>
      </history>
      <copyright-statement>©Xiaoyue Feng, Hao Zhang, Yijie Ren, Penghui Shang, Yi Zhu, Yanchun Liang, Renchu Guan, Dong Xu. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 24.05.2019.</copyright-statement>
      <copyright-year>2019</copyright-year>
      <license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/">
        <p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on http://www.jmir.org/, as well as this copyright and license information must be included.</p>
      </license>
      <self-uri xlink:href="http://www.jmir.org/2019/5/e12957/" xlink:type="simple"/>
      <abstract>
        <sec sec-type="background">
          <title>Background</title>
          <p>It is of great importance for researchers to publish research results in high-quality journals. However, it is often challenging to choose the most suitable publication venue, given the exponential growth of journals and conferences. Although recommender systems have achieved success in promoting movies, music, and products, very few studies have explored recommendation of publication venues, especially for biomedical research. No recommender system exists that can specifically recommend journals in PubMed, the largest collection of biomedical literature.</p>
        </sec>
        <sec sec-type="objective">
          <title>Objective</title>
          <p>We aimed to propose a publication recommender system, named Pubmender, to suggest suitable PubMed journals based on a paper’s abstract.</p>
        </sec>
        <sec sec-type="methods">
          <title>Methods</title>
          <p>In Pubmender, pretrained word2vec was first used to construct the start-up feature space. Subsequently, a deep convolutional neural network was constructed to achieve a high-level representation of abstracts, and a fully connected softmax model was adopted to recommend the best journals.</p>
        </sec>
        <sec sec-type="results">
          <title>Results</title>
          <p>We collected 880,165 papers from 1130 journals in PubMed Central and extracted abstracts from these papers as an empirical dataset. We compared different recommendation models such as Cavnar-Trenkle on the Microsoft Academic Search (MAS) engine, a collaborative filtering–based recommender system for the digital library of the Association for Computing Machinery (ACM) and CiteSeer. We found the accuracy of our system for the top 10 recommendations to be 87.0%, 22.9%, and 196.0% higher than that of MAS, ACM, and CiteSeer, respectively. In addition, we compared our system with Journal Finder and Journal Suggester, which are tools of Elsevier and Springer, respectively, that help authors find suitable journals in their series. The results revealed that the accuracy of our system was 329% higher than that of Journal Finder and 406% higher than that of Journal Suggester for the top 10 recommendations. Our web service is freely available at https://www.keaml.cn:8081/.</p>
        </sec>
        <sec sec-type="conclusions">
          <title>Conclusions</title>
          <p>Our deep learning–based recommender system can suggest an appropriate journal list to help biomedical scientists and clinicians choose suitable venues for their papers.</p>
        </sec>
      </abstract>
      <kwd-group>
        <kwd>recommender system</kwd>
        <kwd>deep learning</kwd>
        <kwd>convolutional neural network</kwd>
        <kwd>biomedical literature</kwd>
        <kwd>PubMed</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec sec-type="introduction">
      <title>Introduction</title>
      <sec>
        <title>Background</title>
        <p>With the fast-growing research activities, more biomedical papers are being published in thousands of journals worldwide. For example, PubMed Central (PMC) has 5.2 million papers and 7409 journals covering biomedical and life sciences [<xref ref-type="bibr" rid="ref1">1</xref>]. Although these publications play a major role in disseminating research outcome, the growth of journal publications imposes a challenge for selections of appropriate publication venues. It is vital that authors submit to the right journal that meets the journal scope and provides sound reviews. It is equally important that they reach their intended audience and obtain a large number of citations [<xref ref-type="bibr" rid="ref2">2</xref>]. However, researchers are unfamiliar with all the journals related to their work for choosing the most suitable one for submitting a paper. Moreover, different publication scopes of journals and research interests of reviewers and editors may affect the decision of a submitted manuscript. If the submitted paper cannot meet the interests of a publication venue and its editors and reviewers, it may lead to rejection, delay, or less readership. An appropriate recommender system can help solve this problem.</p>
        <p>Recommender systems have been proven to serve as an effective method for decision making in many areas such as music, movies, and information media choices [<xref ref-type="bibr" rid="ref3">3</xref>-<xref ref-type="bibr" rid="ref6">6</xref>]. The well-known techniques of recommender systems are content-based recommendation [<xref ref-type="bibr" rid="ref7">7</xref>,<xref ref-type="bibr" rid="ref8">8</xref>], collaborative filtering recommendation [<xref ref-type="bibr" rid="ref4">4</xref>,<xref ref-type="bibr" rid="ref9">9</xref>], and hybrid recommendation [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref10">10</xref>]. Content-based recommender systems recommend an item to a user based on a description of the item. Collaborative filtering methods and hybrid methods may outperform the content-based recommendations by applying user data, if available. However, after the user privacy issue of Facebook in 2018 and the introduction of European Union General Data Protection Regulation, user data are no longer easy to obtain. Moreover, in many domains, especially in material recommendation, there are no user data available for collaborative filtering methods at the beginning [<xref ref-type="bibr" rid="ref11">11</xref>], which is regarded as a cold-start problem. Content-based recommendations do not need any user information and are more suitable for solving these problems [<xref ref-type="bibr" rid="ref12">12</xref>].</p>
        <p>Based on the content-based recommendation strategy, several attempts have been made to create recommender systems for medical applications and scientific literature. Using geotagged mobile search logs, Agarwal et al [<xref ref-type="bibr" rid="ref13">13</xref>] adopted a Random Forest model to predict medical visits. Using topic, writing style, author information, citation information, abstract, and title as information items, latent Dirichlet allocation [<xref ref-type="bibr" rid="ref14">14</xref>] and k-nearest neighbor [<xref ref-type="bibr" rid="ref15">15</xref>] were used to classify the scientific literature for recommendation [<xref ref-type="bibr" rid="ref2">2</xref>,<xref ref-type="bibr" rid="ref12">12</xref>,<xref ref-type="bibr" rid="ref16">16</xref>,<xref ref-type="bibr" rid="ref17">17</xref>]. Luong et al [<xref ref-type="bibr" rid="ref18">18</xref>] used the coauthors’ network as advanced information to recommend a publication venue. Beel et al [<xref ref-type="bibr" rid="ref19">19</xref>] conducted a literature survey on recommender systems, exploring their methods, evaluation measurements, and datasets. For most of these recommender systems, the high-dimensional and sparse matrix computation is a critical problem [<xref ref-type="bibr" rid="ref20">20</xref>].</p>
        <p>Because of the mismatches caused by ambiguity in text comparisons, the content-based recommendation approach may cause a high error rate [<xref ref-type="bibr" rid="ref21">21</xref>]. Recently, due to the ability of discovering intricate structures and deep semantics in high-dimensional data, deep learning methods have succeeded in many areas and recently been proposed to build recommender systems for both collaborative filtering and content-based approaches. Hinton et al [<xref ref-type="bibr" rid="ref22">22</xref>] proposed restricted Boltzmann machines for modeling tabular or count data as a collaborative filtering model on the Netflix data set. McAuley et al [<xref ref-type="bibr" rid="ref23">23</xref>] proposed an image-based recommendation, which adopted a deep learning model to extract image features. Van den Oord et al [<xref ref-type="bibr" rid="ref24">24</xref>] applied a deep convolutional neural network (CNN) to predict latent factors from music audios for music recommendation. Wang et al [<xref ref-type="bibr" rid="ref11">11</xref>] proposed a collaborative deep learning model to jointly perform deep representation learning for the content information and collaborative filtering of a rating matrix. However, to the best of our knowledge, these deep learning techniques have not been used in any biomedical literature recommender system.</p>
        <p>Most current venue recommendation studies focus on computer science and technology, but not on the biomedical field. Biomedical sciences are highly interdisciplinary and often link to engineering, medicine, biology, physics, psychology, etc, thereby serving more journals and more diverse topics than any other field. Hence, the development of a recommender system is more essential and challenging for the biomedical sciences than any other discipline. Furthermore, previous recommender systems were based on shallow machine learning methods and social networks. They were generally keyword-based methods and did not take semantics into account. In addition, the few existing systems only focus on journals under a certain organization, such as Elsevier, IEEE, and Springer, instead of PubMed.</p>
      </sec>
      <sec>
        <title>Aim</title>
        <p>In contrast to our previous study on computer science publication recommendations using conventional machine learning approaches [<xref ref-type="bibr" rid="ref12">12</xref>], we proposed a deep learning–based recommender system for biomedical publication venues, named Pubmender. Due to the copious vocabulary of biomedical literature, the traditional vector space model can lead to high-dimensional and sparse problems. To address this issue, dimensionality reduction methods are needed before learning the pattern. Moreover, initializing text matrix by pretrained word embedding is more beneficial for training neural networks than random initialized embedding [<xref ref-type="bibr" rid="ref25">25</xref>]. Accordingly, we applied a word2vec model for our study instead of using the conventional vector space model employed in our previous publication venue recommender system. In addition, deep learning models are able to learn multiple-level abstract representations of data with syntactic and semantic information, since more abstract concepts can be constructed with multiple processing layers [<xref ref-type="bibr" rid="ref26">26</xref>]. We applied the deep learning approach to provide recommendations of journals for biomedical researchers. Unlike shallow learning, the state-of-the-art embedding method and deep CNN in Pubmender were trained from 837,882 papers in 1130 biomedical journals. This method can help researchers find a variety of choices, without being limited to their own knowledge of journals.</p>
      </sec>
    </sec>
    <sec sec-type="methods">
      <title>Methods</title>
      <sec>
        <title>Pubmender System</title>
        <p><xref ref-type="fig" rid="figure1">Figure 1</xref> shows the architecture and workflow of our Pubmender system. It consists of user interface, data preprocessing, abstract representation, classification, and ranking phase.</p>
        <p>The user interface obtains the input data (an abstract submitted by a user) and presents the recommendation results to the user. The data acquirement is followed by data preprocessing and information extraction. At the start of our deep learning model, the abstract representation phase converts an abstract to a vector. The original abstract vector is a concatenation of pretrained word vectors. Subsequently, deep CNN is applied to train the model to achieve high-level abstract representation. A three-layer fully connected network with a softmax operation is applied to classify papers based on the obtained abstract vectors. The recommendation list of the top N journals obtained from the ranking phase is presented to the user.</p>
      </sec>
      <sec>
        <title>Data Preprocessing Methods</title>
        <p>The data were downloaded from the File Transfer Protocol service of PubMed Central (PMC) [<xref ref-type="bibr" rid="ref27">27</xref>], containing 1,534,649 papers. Based on the journal list of PMC, we selected normal journals deposited under full participation or the US National Institutes of Health portfolio mode, excluding records labeled “Predecessor,” “No New Content,” and “Now Select.” Papers from Jan 2007 to Apr 2017 were selected. Papers with no abstracts or with fewer than 200 characters in abstracts were deleted. Journals containing fewer than 100 papers were also removed. Finally, 880,165 papers in the XML format from 1130 journals were used in our study.</p>
        <p>Each PMC file is a semistructured XML document and contains various tags, such as &lt;title&gt;, &lt;abstract&gt;, and &lt;issn&gt;. We extracted the content in &lt;abstract&gt;, &lt;ISSN&gt;, and &lt;pub-date&gt; fields from the raw XML files. Then, pissn and eissn in the ISSN field were replaced by “LocatorPlus ID,” which is the unique identification for a journal in the US National Library of Medicine catalog. After extraction, each abstract was stored in a corresponding file. Natural Language Toolkit was adopted to operate word segmentation [<xref ref-type="bibr" rid="ref28">28</xref>].</p>
      </sec>
      <sec>
        <title>Abstract Representation</title>
        <p>In Pubmender, the recommendation task is formulated into a multilabel classification problem, where the text representation and classification methods are critical. For abstracts, we originally embedded abstracts with pretrained word vectors. Thereafter, the original embeddings were fed into CNN to achieve more abstract representation as explained below.</p>
        <fig id="figure1" position="float">
          <label>Figure 1</label>
          <caption>
            <p>Architecture of our Pubmender system. CNN: convolutional neural network; ISSN: International Standard Serial Number.</p>
          </caption>
          <graphic xlink:href="jmir_v21i5e12957_fig1.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <p>Let <bold>v</bold><sub><italic>i</italic> </sub>∈R<sup><italic>k</italic> </sup> be the <italic>k</italic>-dimensional word vector corresponding to the <italic>i</italic>-th word in the abstract <italic>A</italic>. An original representation of <italic>A</italic> is represented as a matrix <bold>V</bold>={<bold>v</bold><sub>1</sub>,…,<bold>v</bold><sub><italic>m</italic></sub>}<sup>T</sup>, which is the concatenation of the words’ vectors. Due to the different sizes of abstracts, we set <italic>m</italic> as the maximum count of words in an abstract. A padding operation with zeros was adopted for input with fewer than <italic>m</italic> words in an abstract and a tail truncation operation for more than <italic>m</italic> words. The vectors of words adopt pretrained vectors using word embedding and are induced from the PubMed abstracts and PubMed Central full text. The word2vec tool [<xref ref-type="bibr" rid="ref29">29</xref>] was adopted for word embedding using the skip-gram model with a window size of <italic>h</italic>, hierarchical softmax training, and a frequent word subsampling threshold <inline-graphic xlink:href="jmir_v21i5e12957_fig21.png" mimetype="image" xlink:type="simple"/>to create <italic>k</italic>-dimensional vectors. Word vectors are initialized by zeros if they are not in the pretrained vocabulary. Finally, the representation of an abstract is matrix <bold>V</bold> with a dimensionality of <italic>m*k</italic>. It was used as the input to feed to the next step. To achieve more abstract and semantic features, we adopted CNN to extract semantic information.</p>
        <p><xref ref-type="fig" rid="figure2">Figure 2</xref> shows the structure of our deep CNN model. There are three convolutional and max-pooling layers in CNN, one fully connected layer, one hidden layer, and one softmax layer for classification. For an abstract, <italic>A</italic>(<italic>w</italic><sub>1</sub>,<italic>w</italic><sub>2</sub>,...,<italic>w<sub>n</sub></italic>) with <italic>w<sub>i</sub></italic> represents the <italic>i-</italic>th word and <bold>v</bold><sub><italic>i</italic></sub>∈R<sup><italic>k</italic></sup> is the <italic>k</italic>-dimensional word vector corresponding to word <italic>w<sub>i</sub></italic>. The abstract is represented as <bold>v</bold><sub>1:<italic>m</italic></sub>=<bold>v</bold><sub>1</sub>⨁<bold>v</bold><sub>2</sub>⨁...⨁<bold>v</bold><sub><italic>m</italic></sub> (1), where ⨁ is the concatenation operator, <italic>m</italic> is the maximum length of abstracts (a scalar), and <bold>v</bold><italic><sub>i:i</sub><sub>+</sub><sub>j</sub></italic> refers to the vector of concatenation of the words <italic>w<sub>i</sub></italic>,<italic>w<sub>i</sub></italic><sub>+1</sub>,…,<italic>w<sub>i</sub></italic><sub>+</sub><sub><italic>j</italic></sub>. The first convolutional layer performs as a one-dimensional convolution operation on sliding windows of <italic>h</italic><sub>1</sub> words to produce a phrase feature. For example, a feature <italic>c<sub>ji</sub></italic> is generated from a window of words <bold>v</bold><sub><italic>i:i+h</italic><sub>1</sub>–1</sub> by <italic>c<sub>ji</sub></italic>=<italic>g(f<sub>j</sub></italic>▪<bold>v</bold><sub><italic>i:i+h</italic><sub>1</sub>–1</sub>+<bold>b</bold><sub>1</sub>) (2). Here, <bold>b</bold><sub>1</sub>∈R is a bias term and <italic>g</italic> is a nonlinear function such as rectified linear unit (ReLu). <italic>f<sub>j</sub></italic>∈R<sup><italic>k</italic> × <italic>h</italic><sub>1</sub></sup> is the <italic>j</italic>-th convolutional kernel, whose shape is <italic>k</italic> × <italic>h</italic><sub>1</sub>, where <italic>k</italic> is the dimension of word vectors and <italic>h</italic><sub>1</sub> is the window size. This kernel is applied to each possible window of words in the abstract {<bold>v</bold><sub>1:<italic>h</italic><sub>1</sub></sub>,<bold>v</bold><sub>2:<italic>h</italic><sub>1</sub>+1</sub>,...,<bold>v</bold><sub><italic>m–h</italic><sub>1</sub>+1:<italic>m</italic></sub>} to produce a feature map <bold>C</bold><sub><italic>j</italic></sub>=[<italic>c</italic><sub><italic>j</italic>1</sub>, <italic>c</italic><sub><italic>j</italic>2</sub>...,<italic>c</italic><sub><italic>j,m–h</italic><sub>1</sub>+1</sub>] (3) with <bold>C</bold><sub><italic>j</italic></sub>∈R<sup><italic>m–h</italic><sub>1</sub>+1</sup>.</p>
        <p>If there are <italic>r</italic><sub>1</sub> convolutional kernels, then <inline-graphic xlink:href="jmir_v21i5e12957_fig11.png" mimetype="image" xlink:type="simple"/>is the result of the first convolution operation on <bold>V</bold>. The pooling operation is then carried out on <bold>C<sup>(1)</sup></bold>. Its function is to progressively reduce the spatial size of the representation to extract the key features and reduce the number of dimensions in the network. The pooling layer operates independently on every depth slice of the input and resizes it spatially, using the max-pooling operation [<xref ref-type="bibr" rid="ref30">30</xref>] in every two-unit window for each <bold>C<sub>j</sub><sup>(1)</sup></bold>. <bold>P<sup>(1)</sup></bold>, described below, is the result of the max-pooling operation: <inline-graphic xlink:href="jmir_v21i5e12957_fig12.png" mimetype="image" xlink:type="simple"/> (4), where <italic>j</italic> is the <italic>j</italic>-th filter of the convolutional operation <inline-graphic xlink:href="jmir_v21i5e12957_fig13.png" mimetype="image" xlink:type="simple"/> (5).</p>
        <p>The second and third convolutional and pooling layers work the same as Equations (2) and (5). Following the three convolutional and pooling operations is the fully connected layer. Here, the input is represented with a more abstract feature <inline-graphic xlink:href="jmir_v21i5e12957_fig14.png" mimetype="image" xlink:type="simple"/>, where <italic>r</italic><sub>3</sub> is the number of third-layer convolutional filters. The three convolutional and pooling operations indicate a phrase-level feature, a sentence-level feature, and an abstract-level feature. The algorithm of abstract embedding is listed in <xref ref-type="boxed-text" rid="box1">Textbox 1</xref>.</p>
        <fig id="figure2" position="float">
          <label>Figure 2</label>
          <caption>
            <p>The structure of our deep convolutional neural network model.</p>
          </caption>
          <graphic xlink:href="jmir_v21i5e12957_fig2.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/>
        </fig>
        <boxed-text id="box1" position="float">
          <title>Abstract embedding algorithm.</title>
          <p><bold>Input:</bold></p>
          <list list-type="bullet">
            <list-item>
              <p>Embedding the abstract <italic>A</italic> to matrix <bold>V</bold>={<bold>v</bold><sub>1</sub>,…,<bold>v</bold><italic><sub>m</sub></italic>}<sup>T</sup></p>
            </list-item>
            <list-item>
              <p><italic>r<sub>t</sub></italic> is the number of convolutional filters of layer <italic>t</italic>, where <italic>t</italic>=1, 2, 3</p>
            </list-item>
            <list-item>
              <p><italic>h<sub>t</sub></italic> is the convolutional window size of layer <italic>t</italic>, where <italic>t</italic>=1, 2, 3</p>
            </list-item>
          </list>
          <p><bold>Output:</bold></p>
          <p><inline-graphic xlink:href="jmir_v21i5e12957_fig14.png" mimetype="image" xlink:type="simple"/></p>
          <p><bold>Procedure:</bold></p>
          <list list-type="order">
            <list-item>
              <p><inline-graphic xlink:href="jmir_v21i5e12957_fig15.png" mimetype="image" xlink:type="simple"/></p>
            </list-item>
            <list-item>
              <p>for <italic>t</italic>=1, 2, 3</p>
            </list-item>
            <list-item>
              <p>for <italic>j</italic>=1, 2,…<italic>r<sub>t</sub></italic></p>
            </list-item>
            <list-item>
              <p>for <italic>i</italic>=1, 2,…<italic>m–h</italic><sub>t</sub>+1</p>
            </list-item>
            <list-item>
              <p><inline-graphic xlink:href="jmir_v21i5e12957_fig22.png" mimetype="image" xlink:type="simple"/></p>
            </list-item>
            <list-item>
              <p>End for</p>
            </list-item>
            <list-item>
              <p>for <italic>i</italic>=1, 2,… (<italic>m</italic>–<italic>h<sub>t</sub></italic>+1)/2</p>
            </list-item>
            <list-item>
              <p><inline-graphic xlink:href="jmir_v21i5e12957_fig24.png" mimetype="image" xlink:type="simple"/></p>
            </list-item>
            <list-item>
              <p>End for</p>
            </list-item>
            <list-item>
              <p><inline-graphic xlink:href="jmir_v21i5e12957_fig23.png" mimetype="image" xlink:type="simple"/></p>
            </list-item>
            <list-item>
              <p>End for</p>
            </list-item>
            <list-item>
              <p><inline-graphic xlink:href="jmir_v21i5e12957_fig16.png" mimetype="image" xlink:type="simple"/></p>
            </list-item>
            <list-item>
              <p>End for</p>
            </list-item>
          </list>
        </boxed-text>
      </sec>
      <sec>
        <title>Softmax Classification</title>
        <p>A fully connected softmax layer is the last layer of Pubmender. Given the training sample, <italic>A</italic>, where <italic>T</italic> is the number of possible labels, <italic>z</italic> is the class score for the sample, and the estimated probabilities <italic>S<sub>j</sub></italic>∈[0,1) for each label j∈{1,2…<italic>T</italic>} the softmax formula is: <graphic xlink:href="jmir_v21i5e12957_fig17.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/></p>
        <p>We trained the entire model by minimizing the cross-entropy error defined as <inline-graphic xlink:href="jmir_v21i5e12957_fig18.png" mimetype="image" xlink:type="simple"/>(7), where <italic>Y</italic> is the true classification output. This is a one-hot encoding of size <italic>T</italic>, where all elements except one are 0, and one element is 1. This element marks the correct class for the data classified. We employed the optimizer Adam to learn the model parameters, which is a variant of stochastic gradient descent [<xref ref-type="bibr" rid="ref31">31</xref>].</p>
      </sec>
    </sec>
    <sec sec-type="results">
      <title>Results</title>
      <sec>
        <title>Datasets</title>
        <p>After data preprocessing, there were 880,165 preprocessed papers from PMC in 1130 open-access journals from Jan 2007 to Apr 2017. The “LocatorPlus ID” assigned to each journal by PMC is regarded as the classification label of a paper. We generated four data sets based on these papers. The first data set included all papers from 2007 to 2016, which was used to choose the feature representation method and train the prediction models. Papers in 2017 formed the second data set, which was used as the test set to verify Pubmender’s performance. The last two datasets chose papers from publications in Elsevier and Springer from 2017, which were used to compare our Pubmender with Journal Finder and Suggester. The statistics of the first dataset are described in <xref ref-type="table" rid="table1">Table 1</xref>.</p>
        <p>One of the biggest challenges of these datasets is that the data distribution is highly imbalanced. In the first dataset, 60 journals published more than 2000 papers, while 740 journals published fewer than 400 papers. The number of papers in “PLOS One” was 153,608, which is larger than the number in other journals, based on its extensive and comprehensive scope. “Scientific Reports” ranked second, with 37,864 papers published, while “Horticulture Research” only published 100 papers. The average paper count was 741, and 934 journals had fewer than that number.</p>
        <table-wrap position="float" id="table1">
          <label>Table 1</label>
          <caption>
            <p>Details of the first dataset (Jan 2007 to Dec 2016).</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="30"/>
            <col width="550"/>
            <col width="280"/>
            <col width="140"/>
            <thead>
              <tr valign="bottom">
                <td colspan="2">Statistic</td>
                <td>Number of journals<sup>a</sup></td>
                <td>Number of papers<sup>b</sup></td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td colspan="2"><bold>Size</bold></td>
                <td> </td>
                <td> </td>
              </tr>
              <tr valign="top">
                <td> </td>
                <td>100≤x<sup>c</sup>≤400</td>
                <td>740</td>
                <td>157,038</td>
              </tr>
              <tr valign="top">
                <td> </td>
                <td>400&lt;x≤2000</td>
                <td>330</td>
                <td>259,676</td>
              </tr>
              <tr valign="top">
                <td> </td>
                <td>2000&lt;x≤10,000</td>
                <td>55</td>
                <td>195,426</td>
              </tr>
              <tr valign="top">
                <td> </td>
                <td>&gt;10,000</td>
                <td>5</td>
                <td>225,742</td>
              </tr>
              <tr valign="top">
                <td> </td>
                <td>Total</td>
                <td>1130</td>
                <td>837,882</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Maximum class size</td>
                <td>1</td>
                <td>153,608</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Minimum class size</td>
                <td>4</td>
                <td>100</td>
              </tr>
              <tr valign="top">
                <td colspan="2">Average class size</td>
                <td>N/A<sup>d</sup></td>
                <td>741</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table1fn1">
              <p><sup>a</sup>This represents the total number of journals in this range.</p>
            </fn>
            <fn id="table1fn2">
              <p><sup>b</sup>This represents the total number of papers published in all journals in this range.</p>
            </fn>
            <fn id="table1fn3">
              <p><sup>c</sup>x represents the number of papers published in one journal.</p>
            </fn>
            <fn id="table1fn4">
              <p><sup>d</sup>N/A: not applicable.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
      </sec>
      <sec>
        <title>Parameters and Measurements</title>
        <p>For CNN, three convolutional and three pooling operations were adopted. The pretrained word vectors generated by word2vec, available from Evex [<xref ref-type="bibr" rid="ref29">29</xref>], were used. The window size <italic>h</italic> was 5 and threshold <inline-graphic xlink:href="jmir_v21i5e12957_fig21.png" mimetype="image" xlink:type="simple"/>was 0.001. The dimension of a pretrained vector was 200, that is, <italic>k</italic>=200. The length of abstract had a fixed size <italic>m</italic>, which is the maximum number of the words that most abstracts contain. <xref ref-type="table" rid="table2">Table 2</xref> shows the word statistic details of the papers. With the statistics, only 43,328 of 837,882 papers (5%) contained abstracts with more than 350 words. Therefore, we chose <italic>m</italic>=350. A zero-padding operation was applied for abstracts with fewer than 350 words, together with a tail-truncation operation for abstracts containing more than 350 words.</p>
        
      
        <p>The convolutional operation parameters are listed in <xref ref-type="table" rid="table3">Table 3</xref>. The activation function adopted the ReLU function. In the pooling layer, the size of max-pooling filters was two, applied with a stride of two down samples. The parameters of the following layers (pooling layers) had the same parameter settings. Normalization and dropout strategies were used in the fully connected layer. Rate of dropout was 0.2 and L<sub>2</sub> normalization was adopted.</p>
      </sec>
      <sec>
        <title>Evaluation of Recommendation Results</title>
        <sec>
          <title>Toy Experiment</title>
          <p>We designed a toy experiment to validate the deep learning method. In the first dataset, 421,168 papers were chosen from 60 journals with more than 2000 papers. The training set and test set contained 37,951 (90%) and 4,217 (10%) papers, respectively. We selected bi-directional long short-term memory (Bi-LSTM) and fastText [<xref ref-type="bibr" rid="ref32">32</xref>] as comparison models for Pubmender. Bi-LSTM represents the recurrent neural network model with the max-pooling operation from a previous study [<xref ref-type="bibr" rid="ref33">33</xref>]. Pretrained word vectors, generated by word2vec from a previous study [<xref ref-type="bibr" rid="ref29">29</xref>], were used as the input original word vectors for fastText and Pubmender.</p>
          <p>To evaluate the performance of our system, top-<italic>N</italic> accuracy was adopted as a measurement, which is defined as the probability that the expected label is in the top <italic>N</italic> predicted classes. For top-<italic>N</italic>, if the journal containing the abstract is among the top <italic>N</italic> ranked journals, the classification is correct. The symbol acc@<italic>N</italic> represents the accuracy of top-<italic>N</italic>, <italic>N</italic>=1, <italic>N=</italic>3, and <italic>N</italic>=5. The comparison of accuracy is listed in <xref ref-type="table" rid="table4">Table 4</xref>, which shows that both deep learning approaches outperformed fastText in all the three measurements. The accuracy of Bi-LSTM is nearly the same as that of Pubmender. However, the running time of Pubmender was 2660 abstracts per second, which is 78% faster than Bi-LSTM (1495 abstracts/second), and Bi-LSTM needs more memory.</p>
            <table-wrap position="float" id="table2">
          <label>Table 2</label>
          <caption>
            <p>Word statistics of abstracts.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="720"/>
            <col width="280"/>
            <thead>
              <tr valign="top">
                <td>Size</td>
                <td>Number of abstracts</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>20≤x<sup>a</sup>&lt;50</td>
                <td>25,499</td>
              </tr>
              <tr valign="top">
                <td>50≤x&lt;100</td>
                <td>76,614</td>
              </tr>
              <tr valign="top">
                <td>100≤x&lt;150</td>
                <td>139,420</td>
              </tr>
              <tr valign="top">
                <td>150≤x&lt;200</td>
                <td>227,993</td>
              </tr>
              <tr valign="top">
                <td>200≤x&lt;250</td>
                <td>191,156</td>
              </tr>
              <tr valign="top">
                <td>250≤x&lt;300</td>
                <td>87,597</td>
              </tr>
              <tr valign="top">
                <td>300≤x&lt;350</td>
                <td>46,275</td>
              </tr>
              <tr valign="top">
                <td>x&gt;350</td>
                <td>43,328</td>
              </tr>
            </tbody>
          </table>
          <table-wrap-foot>
            <fn id="table2fn1">
              <p><sup>a</sup>x denotes the number of words in the abstract.</p>
            </fn>
          </table-wrap-foot>
        </table-wrap>
       
          <table-wrap position="float" id="table3">
            <label>Table 3</label>
            <caption>
              <p>Hyperparameters of convolutional operation.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="580"/>
              <col width="280"/>
              <col width="140"/>
              <thead>
                <tr valign="top">
                  <td>Convolutional layer</td>
                  <td>Convolution kernel count</td>
                  <td>Window size</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>First</td>
                  <td>256</td>
                  <td>3</td>
                </tr>
                <tr valign="top">
                  <td>Second</td>
                  <td>128</td>
                  <td>4</td>
                </tr>
                <tr valign="top">
                  <td>Third</td>
                  <td>96</td>
                  <td>5</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
          <table-wrap position="float" id="table4">
            <label>Table 4</label>
            <caption>
              <p>Accuracy of Bi-LSTM, fastText, and Pubmender. Italicized values indicate the best results. acc@N represents the accuracy for top-N selection.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="580"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <thead>
                <tr valign="top">
                  <td>Methods</td>
                  <td>acc@1</td>
                  <td>acc@3</td>
                  <td>acc@5</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>fastText</td>
                  <td>0.66</td>
                  <td>0.86</td>
                  <td>0.92</td>
                </tr>
                <tr valign="top">
                  <td>Bi-LSTM<sup>a</sup> (max-pooling)</td>
                  <td>0.71</td>
                  <td>0.90</td>
                  <td>0.95</td>
                </tr>
                <tr valign="top">
                  <td>Pubmender</td>
                  <td><italic>0.72</italic></td>
                  <td><italic>0.92</italic></td>
                  <td><italic>0.96</italic></td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table4fn1">
                <p><sup>a</sup>Bi-LSTM<bold>:</bold> bi-directional long short-term memory.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>

           </sec>
        <sec>
          <title>First Dataset Result</title>
          <p>For the first dataset, there were 837,882 papers in 1130 journals from 2007 to 2016. The training, validation, and test sets contained 670,306 (80%), 83,788 (10%), and 83,788 (10%) randomly selected papers, respectively. The results of and comparisons with previous work are provided in <xref ref-type="table" rid="table5">Table 5</xref>.</p>
          <p>The other two systems were from three widely used digital libraries: Association for Computing Machinery (ACM) [<xref ref-type="bibr" rid="ref16">16</xref>], CiteSeer [<xref ref-type="bibr" rid="ref16">16</xref>], and Microsoft Academic Search (MAS) [<xref ref-type="bibr" rid="ref2">2</xref>]. The results from <xref ref-type="table" rid="table5">Table 5</xref> show that Pubmender achieved the best performance. The proposed system can achieve 0.50 on acc@1 and 0.86 on acc@10. Our system improved performance by 225% over CiteSeer and MAS in terms of acc@5, and by 87% over MAS and 196% over CiteSeer in terms of acc@10. The system described by Yang and Davidson [<xref ref-type="bibr" rid="ref16">16</xref>] used topic and writing-style information, and the system described by Medvet et al [<xref ref-type="bibr" rid="ref2">2</xref>] used the abstracts and titles. However, our Pubmender obtained the best accuracy by using abstracts only.</p>
          <p>To present the ability of handling imbalanced data of Pubmender, we divided the test set into four classes (tiny, small, medium, and large) according to the paper counts of different journals. From <xref ref-type="table" rid="table6">Table 6</xref>, for the tiny set, Pubmender achieved 0.27 accuracy on the acc@1 and 0.54 on acc@5, which are greater than the accuracy on acc@5 and acc@10 (in <xref ref-type="table" rid="table5">Table 5</xref>) from MAS and CiteSeer, respectively. The accuracy of the top-10 (acc@10) of a large set reached 0.98. In the paper by Medvet et al [<xref ref-type="bibr" rid="ref2">2</xref>], 58,466 papers were partitioned almost uniformly into 300 conferences from the MAS. In CiteSeer [<xref ref-type="bibr" rid="ref16">16</xref>], 35,020 selected papers were published across 739 venues, each of which had at least 20 papers. The average number of papers for each venue was 47. Therefore, the CiteSeer dataset is almost balanced. In contrast, the imbalance of our data, as shown in <xref ref-type="table" rid="table1">Table 1</xref>, was very critical. The sizes of classes in our dataset ranged from 100 to 153,608 papers; for example, the number of papers in “PLOS One” was 153,608, which is 270 times the average number of papers in all journals. Compared with balanced data, the classification of critically imbalanced data was a complex problem to tackle. For this problem, our model achieved satisfactory results.</p>
                    <table-wrap position="float" id="table5">
            <label>Table 5</label>
            <caption>
              <p>Accuracy of the classification by Pubmender and other systems. Italicized values indicate the best results. acc@N represents the accuracy for top-N selection.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="160"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <thead>
                <tr valign="top">
                  <td>Methods</td>
                  <td>Paper count</td>
                  <td>Journal count</td>
                  <td>acc@1</td>
                  <td>acc@3</td>
                  <td>acc@5</td>
                  <td>acc@10</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Pubmender</td>
                  <td>837,882</td>
                  <td>1130</td>
                  <td><italic>0.50</italic></td>
                  <td><italic>0.71</italic></td>
                  <td><italic>0.78</italic></td>
                  <td><italic>0.86</italic></td>
                </tr>
                <tr valign="top">
                  <td>MAS<sup>a</sup> [<xref ref-type="bibr" rid="ref2">2</xref>]</td>
                  <td>58,466</td>
                  <td>300</td>
                  <td>—<sup>b</sup></td>
                  <td>—</td>
                  <td>0.24</td>
                  <td>0.46</td>
                </tr>
                <tr valign="top">
                  <td>ACM<sup>c</sup> [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                  <td>172,890</td>
                  <td>2197</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.56</td>
                  <td>0.70</td>
                </tr>
                <tr valign="top">
                  <td>CiteSeer [<xref ref-type="bibr" rid="ref16">16</xref>]</td>
                  <td>35,020</td>
                  <td>739</td>
                  <td>—</td>
                  <td>—</td>
                  <td>0.24</td>
                  <td>0.29</td>
                </tr>
              </tbody>
            </table>
            <table-wrap-foot>
              <fn id="table5fn1">
                <p><sup>a</sup>MAS: Microsoft Academic Search.</p>
              </fn>
              <fn id="table5fn2">
                <p><sup>b</sup>Experimental evaluation is not available.</p>
              </fn>
              <fn id="table5fn3">
                <p><sup>c</sup>ACM: Association for Computing Machinery.</p>
              </fn>
            </table-wrap-foot>
          </table-wrap>
          <table-wrap position="float" id="table6">
            <label>Table 6</label>
            <caption>
              <p>Pubmender accuracy at top N(@N) of imbalance class data. acc@N represents the accuracy for top-N selection.</p>
            </caption>
            <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
              <col width="300"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <col width="140"/>
              <thead>
                <tr valign="top">
                  <td>Paper count range</td>
                  <td>acc@1</td>
                  <td>acc@3</td>
                  <td>acc@5</td>
                  <td>acc@10</td>
                  <td>Paper count</td>
                </tr>
              </thead>
              <tbody>
                <tr valign="top">
                  <td>Tiny</td>
                  <td>0.27</td>
                  <td>0.44</td>
                  <td>0.54</td>
                  <td>0.66</td>
                  <td>16,337</td>
                </tr>
                <tr valign="top">
                  <td>Small</td>
                  <td>0.43</td>
                  <td>0.63</td>
                  <td>0.72</td>
                  <td>0.82</td>
                  <td>26,259</td>
                </tr>
                <tr valign="top">
                  <td>medium</td>
                  <td>0.62</td>
                  <td>0.81</td>
                  <td>0.88</td>
                  <td>0.94</td>
                  <td>19,588</td>
                </tr>
                <tr valign="top">
                  <td>Large</td>
                  <td>0.66</td>
                  <td>0.91</td>
                  <td>0.96</td>
                  <td>0.98</td>
                  <td>22,579</td>
                </tr>
                <tr valign="top">
                  <td>All</td>
                  <td>0.50</td>
                  <td>0.71</td>
                  <td>0.78</td>
                  <td>0.86</td>
                  <td>84,763</td>
                </tr>
              </tbody>
            </table>
          </table-wrap>
          <p>Moreover, excluding accuracy, we choose precision, recall, and the F1-score as measurements. For an individual class <bold>C</bold><sub><italic>i</italic> </sub>, the assessment is defined by <italic>tp<sub>i</sub></italic> (true positives), <italic>fp<sub>i</sub></italic> (false positives), <italic>fn<sub>i</sub></italic> (false negatives), and <italic>tn<sub>i</sub></italic> (true negatives). Accuracy, precision, and recall are calculated from the counts for <bold>C</bold><sub><italic>i</italic> </sub>. Quality of the overall classification is evaluated in two ways: macro-averaging and micro-averaging. The macro-average is the average of the same measures calculated for all classes. With the sum of counts to obtain cumulative <italic>tp</italic>, <italic>fp</italic>, <italic>tn</italic>, and <italic>fn</italic>, micro-average metrics are calculated [<xref ref-type="bibr" rid="ref34">34</xref>]. The following equations show how the desired results are individually achieved: <graphic xlink:href="jmir_v21i5e12957_fig19.png" alt-version="no" mimetype="image" position="float" xlink:type="simple"/></p>
          <p>We listed macro-average and micro-average metrics in <xref ref-type="table" rid="table7">Table 7</xref>. Macro-averaging treats all classes equally, while micro-averaging favors bigger classes. From <xref ref-type="table" rid="table7">Table 7</xref>, it can be seen that when the number of recommended journals increases, the probability of capturing the real journal also increases. Therefore, the recall is increased step by step from top-1 to top-10. With the growth in the number of recommended journals, the number of falsely selected journals is also growing, which results in a decrease in precision. The F1-score favors a balanced view.</p>
        </sec>
        <sec>
          <title>New Data Verification</title>
          <p>To show the performance on new data, 42,283 papers from January 2017 to April 2017 were extracted to make further predictions. This comprised 1321 journals, some of which did not appear in the first dataset. These unseen journals increased the difficulty of prediction. The accuracy of Pubmender on the top-1, 3, 5, and 10 was 0.39, 0.61, 0.68, and 0.76, respectively. The accuracies on acc@5 and acc@10 were 183% and 162% higher than those of CiteSeer, respectively. From these results, we conclude that our proposed recommender system achieves a satisfactory result, even for new data that may not belong to the same data distribution.</p>
        </sec>
      </sec>
      <sec>
        <title>Comparison with Journal Finder and Journal Suggester</title>
        <p>Journal Finder is provided by Elsevier for recommending Elsevier journals [<xref ref-type="bibr" rid="ref35">35</xref>]. There are 45 Elsevier journals in our dataset. Five of them with more papers were selected. The paper counts were 582 for <italic>Medicine</italic>, 193 for <italic>Data in Brief</italic>, 124 for <italic>NeuroImage: Clinical,</italic> 117 for <italic>Redox Biology</italic>, and 87 for <italic>Preventive Medicine Reports</italic>.</p>
        <p>Elsevier’s Journal Finder requires input of the title and abstract of submitted paper, and fields of research. The titles and abstracts were extracted from XML files and then fed into Journal Finder. We chose fields of research in “Engineering,” “GeoSciences,” “Life and Health Science,” and “Chemistry.” The results are listed in <xref ref-type="table" rid="table8">Table 8</xref> and show that our system is much better than Journal Finder.</p>

        <table-wrap position="float" id="table7">
          <label>Table 7</label>
          <caption>
            <p>Macro-average and Micro-average metrics for recommendation results.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="160"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <thead>
              <tr valign="top">
                <td>Metrics</td>
                <td colspan="3">Macro-average</td>
                <td colspan="3">Micro-average</td>
              </tr>
              <tr valign="top">
                <td><break/></td>
                <td>Precision</td>
                <td>Recall</td>
                <td>F1</td>
                <td>Precision</td>
                <td>Recall</td>
                <td>F1</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Top-1</td>
                <td>0.38</td>
                <td>0.32</td>
                <td>0.33</td>
                <td>0.50</td>
                <td>0.50</td>
                <td>0.50</td>
              </tr>
              <tr valign="top">
                <td>Top-3</td>
                <td>0.37</td>
                <td>0.50</td>
                <td>0.41</td>
                <td>0.45</td>
                <td>0.71</td>
                <td>0.55</td>
              </tr>
              <tr valign="top">
                <td>Top-5</td>
                <td>0.35</td>
                <td>0.59</td>
                <td>0.42</td>
                <td>0.42</td>
                <td>0.78</td>
                <td>0.55</td>
              </tr>
              <tr valign="top">
                <td>Top-10</td>
                <td>0.32</td>
                <td>0.70</td>
                <td>0.42</td>
                <td>0.38</td>
                <td>0.86</td>
                <td>0.53</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table8">
          <label>Table 8</label>
          <caption>
            <p>Comparison between Pubmender and Journal Finder. Italicized values indicate the best results. acc@N represents the accuracy for top-N selection.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="440"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <thead>
              <tr valign="top">
                <td>Systems</td>
                <td>acc@1</td>
                <td>acc@3</td>
                <td>acc@5</td>
                <td>acc@10</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Pubmender</td>
                <td><italic>0.62</italic></td>
                <td><italic>0.75</italic></td>
                <td><italic>0.84</italic></td>
                <td><italic>0.90</italic></td>
              </tr>
              <tr valign="top">
                <td>Journal Finder</td>
                <td>0.05</td>
                <td>0.12</td>
                <td>0.13</td>
                <td>0.21</td>
              </tr>
              <tr valign="top">
                <td>Improvement (%)</td>
                <td>1140</td>
                <td>525</td>
                <td>546</td>
                <td>329</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
        <table-wrap position="float" id="table9">
          <label>Table 9</label>
          <caption>
            <p>Comparison between Pubmender and Journal Suggester. Italicized values indicate the best results. acc@N represents the accuracy for top-N selection.</p>
          </caption>
          <table width="1000" cellpadding="5" cellspacing="0" border="1" rules="groups" frame="hsides">
            <col width="440"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <col width="140"/>
            <thead>
              <tr valign="top">
                <td>Systems</td>
                <td>acc@1</td>
                <td>acc@3</td>
                <td>acc@5</td>
                <td>acc@10</td>
              </tr>
            </thead>
            <tbody>
              <tr valign="top">
                <td>Pubmender</td>
                <td><italic>0.57</italic></td>
                <td><italic>0.81</italic></td>
                <td><italic>0.87</italic></td>
                <td><italic>0.91</italic></td>
              </tr>
              <tr valign="top">
                <td>Journal Suggester</td>
                <td>0.11</td>
                <td>0.15</td>
                <td>0.17</td>
                <td>0.18</td>
              </tr>
              <tr valign="top">
                <td>Improvement (%)</td>
                <td>418</td>
                <td>440</td>
                <td>412</td>
                <td>406</td>
              </tr>
            </tbody>
          </table>
        </table-wrap>
                <p>Journal Suggester [<xref ref-type="bibr" rid="ref36">36</xref>], recommends journals published by Springer. Journal Suggester also requires input of the title and abstract, and field of research. We chose “Biomedicine” as the field of research. There are 14 journals from Springer in our dataset, and seven of them were chosen for comparison based on a significant number of papers in each journal — <italic>Cell Death &amp; Disease</italic>, <italic>Malaria Journal</italic>, <italic>Nanoscale Research Letters</italic>, <italic>Nature Communications</italic>, <italic>Parasites &amp; Vectors</italic>, <italic>Scientific Reports</italic>, and <italic>Trials</italic>. Each journal chose the top 100 papers according to the size of the XML files. The results are listed in <xref ref-type="table" rid="table9">Table 9</xref>. Again, our system was much better than Journal Suggester.</p>
      </sec>
    </sec>
    <sec sec-type="discussion">
      <title>Discussion</title>
      <p>In this study, Pubmender was proposed to recommend a biomedical publishing venue to user. CNN was used to obtain the abstract representation. Our results show the performance of the system.</p>
      <sec>
        <title>Principal Results</title>
        <p>For biomedical publications, our Pubmender system is the first recommender system with word embedding and deep learning models. It achieves 87.0%, 22.9%, and 196.0% higher accuracy than recommender systems on MAS, ACM, and CiteSeer, respectively. In addition, the experiment results also revealed that the accuracy of our system was superior to that of Journal Finder and Journal Suggester. Our web service is freely available online [<xref ref-type="bibr" rid="ref37">37</xref>].</p>
      </sec>
      <sec>
        <title>Comparison with Prior Work</title>
        <p>Because no paper has been published about biomedical venue recommendations, we cannot perform any exact comparison with previous work. However, some publishers provide tools to help authors choose suitable journals. We chose two tools provided by Elsevier and Springer for comparison. The first one is Journal Finder provided by Elsevier for recommending journals of Elsevier. Pubmender achieved a much higher accuracy than Journal Finder on four metrics. For example, on acc@1, the accuracy of our system reached 0.62 and Journal Finder was given an accuracy rating of 0.05, with 1140% improvement; on acc@10, the accuracy of our system reached 0.84, which is 546% higher than that of Journal Finder. Pubmender also significantly outperformed another tool, Journal Suggester.</p>
      </sec>
      <sec>
        <title>Conclusions</title>
        <p>In this study, we proposed a biomedical publishing venue recommender system—Pubmender. In this system, an abstract is first represented by a vector using the composition of pretrained word vectors. Subsequently, a deep CNN architecture is designed to represent and classify the submitted abstract. The original vectors are converted into more abstract feature vectors containing semantic information using deep CNNs, which overcome the sparse high-dimensional problem.</p>
        <p>The experimental results showed that our proposed system achieves more successful performance than that of MAS, ACM, CiteSeer, Journal Finder, and Journal Suggester. Even for journals containing a small number of abstracts, the performance of Pubmender was satisfactory, because Pubmender’s high-level representation method catches more semantic and structural information from the abstract.</p>
      </sec>
    </sec>
  </body>
  <back>
    <glossary>
      <title>Abbreviations</title>
      <def-list>
        <def-item>
          <term id="abb1">ACM</term>
          <def>
            <p>Association for Computing Machinery</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb2">Bi-LSTM</term>
          <def>
            <p>bi-directional long short-term memory</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb3">CNN</term>
          <def>
            <p>convolutional neural network</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb4">MAS</term>
          <def>
            <p>Microsoft Academic Search</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb5">PMC</term>
          <def>
            <p>PubMed Central</p>
          </def>
        </def-item>
        <def-item>
          <term id="abb6">ReLu</term>
          <def>
            <p>rectified linear unit</p>
          </def>
        </def-item>
      </def-list>
    </glossary>
    <ack>
      <p>This work was partially supported by the US National Institutes of Health (Grant R35-GM126985), the National Natural Science Foundation of China (Grants 61602207 and 61572228), Zhuhai Premier-Discipline Enhancement Scheme (Grant 2015YXXK02), Key Technological Research Projects in Jilin Province (Grant 20190302107GX), and Guangdong Premier Key-Discipline Enhancement Scheme (Grant 2016GDYSZDXK036).</p>
    </ack>
    <fn-group>
      <fn fn-type="con">
        <p>RG, DX, and XF envisioned the study. HZ and PS performed the modeling experiments. YR performed the data preprocessing. YZ and BH built the website. XF drafted the paper; YL, RG, and DX made critical revisions. RG and DX jointly supervised this project as co-corresponding authors. All authors approved the final manuscript.</p>
      </fn>
      <fn fn-type="conflict">
        <p>None declared.</p>
      </fn>
    </fn-group>
    <ref-list>
      <ref id="ref1">
        <label>1</label>
        <nlm-citation citation-type="web">
        <source>PMC: US National Library of Medicine and National Institutes of Health</source>  
        <access-date>2018-11-26</access-date>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://www.ncbi.nlm.nih.gov/pmc/">https://www.ncbi.nlm.nih.gov/pmc/</ext-link>
          <ext-link ext-link-type="webcite" xlink:href="74E8cB679"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref2">
        <label>2</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Medvet</surname>
            <given-names>E</given-names>
          </name>
          <name name-style="western">
            <surname>Bartoli</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Piccinin</surname>
            <given-names>G</given-names>
          </name>
        </person-group>
        <article-title>Publication Venue Recommendation Based on Paper Abstract</article-title>
        <source>Proceedings of the IEEE 26th International Conference on Tools with Artificial Intelligence</source>  
        <year>2014</year>  
        <month>11</month>  
        <conf-name>2014 IEEE 26th International Conference on Tools with Artificial Intelligence</conf-name>
        <conf-date>November 10-12, 2014</conf-date>
        <conf-loc>Limassol, Cyprus</conf-loc>
        <fpage>1004</fpage>  
        <lpage>1010</lpage>  
        <pub-id pub-id-type="doi">10.1109/ICTAI.2014.152</pub-id></nlm-citation>
      </ref>
      <ref id="ref3">
        <label>3</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Bell</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Koren</surname>
            <given-names>Y</given-names>
          </name>
        </person-group>
        <article-title>Scalable Collaborative Filtering with Jointly Derived Neighborhood Interpolation Weights</article-title>
        <source>Proceedings of the Seventh IEEE International Conference on Data Mining 2007</source>  
        <year>2007</year>  
        <conf-name>Seventh IEEE International Conference on Data Mining 2007</conf-name>
        <conf-date>Oct 28-31, 2007</conf-date>
        <conf-loc>Omaha, NE</conf-loc>
        <fpage>43</fpage>  
        <lpage>52</lpage>  
        <pub-id pub-id-type="doi">10.1109/ICDM.2007.90</pub-id></nlm-citation>
      </ref>
      <ref id="ref4">
        <label>4</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Linden</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Smith</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>York</surname>
            <given-names>J</given-names>
          </name>
        </person-group>
        <article-title>Amazon.com recommendations: item-to-item collaborative filtering</article-title>
        <source>IEEE Internet Computing</source>  
        <year>2003</year>  
        <month>01</month>  
        <volume>7</volume>  
        <issue>1</issue>  
        <fpage>76</fpage>  
        <lpage>80</lpage>  
        <pub-id pub-id-type="doi">10.1109/MIC.2003.1167344</pub-id></nlm-citation>
      </ref>
      <ref id="ref5">
        <label>5</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Mobasher</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Dai</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Luo</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Nakagawa</surname>
            <given-names>M</given-names>
          </name>
        </person-group>
        <article-title>Discovery and Evaluation of Aggregate Usage Profiles for Web Personalization</article-title>
        <source>Data Min Knowl Discov</source>  
        <year>2002</year>  
        <month>01</month>  
        <volume>6</volume>  
        <issue>1</issue>  
        <fpage>61</fpage>  
        <lpage>82</lpage>  
        <pub-id pub-id-type="doi">10.1023/A:1013232803866</pub-id></nlm-citation>
      </ref>
      <ref id="ref6">
        <label>6</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Francesco</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Rokach</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Shapira</surname>
            <given-names>B</given-names>
          </name>
        </person-group>
        <article-title>Introduction to Recommender Systems Handbook</article-title>
        <source>Recommender Systems Handbook</source>  
        <year>2019</year>  
        <publisher-loc>Boston, MA</publisher-loc>
        <publisher-name>Springer</publisher-name></nlm-citation>
      </ref>
      <ref id="ref7">
        <label>7</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Brusilovsky</surname>
            <given-names>P</given-names>
          </name>
          <name name-style="western">
            <surname>Kobsa</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Nejdl</surname>
            <given-names>W</given-names>
          </name>
        </person-group>
        <article-title>Content-Based Recommendation Systems</article-title>
        <source>The Adaptive Web: Methods And Strategies Of Web Personalization (lecture Notes In Computer Science)</source>  
        <year>2019</year>  
        <publisher-loc>Berlin, Heidelberg, Germany</publisher-loc>
        <publisher-name>Springer</publisher-name></nlm-citation>
      </ref>
      <ref id="ref8">
        <label>8</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Balabanović</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Shoham</surname>
            <given-names>Y</given-names>
          </name>
        </person-group>
        <article-title>Fab: content-based, collaborative recommendation</article-title>
        <source>Commun ACM</source>  
        <year>1997</year>  
        <month>03</month>  
        <volume>40</volume>  
        <issue>3</issue>  
        <fpage>66</fpage>  
        <lpage>72</lpage>  
        <pub-id pub-id-type="doi">10.1145/245108.245124</pub-id></nlm-citation>
      </ref>
      <ref id="ref9">
        <label>9</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Sarwar</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Karypis</surname>
            <given-names>G</given-names>
          </name>
          <name name-style="western">
            <surname>Konstan</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Riedl</surname>
            <given-names>J</given-names>
          </name>
        </person-group>
        <article-title>Item-based Collaborative Filtering Recommendation Algorithms</article-title>
        <source>Proceedings of the 10th international conference on World Wide Web</source>  
        <year>2001</year>  
        <conf-name>The 10th international conference on World Wide Web</conf-name>
        <conf-date>May 01-05, 2001</conf-date>
        <conf-loc>Hong Kong</conf-loc>
        <fpage>285</fpage>  
        <lpage>295</lpage>  
        <pub-id pub-id-type="doi">10.1145/371920.372071</pub-id></nlm-citation>
      </ref>
      <ref id="ref10">
        <label>10</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Salter</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Antonopoulos</surname>
            <given-names>N</given-names>
          </name>
        </person-group>
        <article-title>CinemaScreen Recommender Agent: Combining Collaborative and Content-Based Filtering</article-title>
        <source>IEEE Intell. Syst</source>  
        <year>2006</year>  
        <month>01</month>  
        <volume>21</volume>  
        <issue>1</issue>  
        <fpage>35</fpage>  
        <lpage>41</lpage>  
        <pub-id pub-id-type="doi">10.1109/MIS.2006.4</pub-id></nlm-citation>
      </ref>
      <ref id="ref11">
        <label>11</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>N</given-names>
          </name>
          <name name-style="western">
            <surname>Yeung</surname>
            <given-names>D</given-names>
          </name>
        </person-group>
        <article-title>Collaborative Deep Learning for Recommender Systems</article-title>
        <source>Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>  
        <year>2015</year>  
        <month>8</month>  
        <conf-name>The 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</conf-name>
        <conf-date>2015 Aug 10-13</conf-date>
        <conf-loc>Sydney, NSW, Australia</conf-loc>
        <publisher-loc>New York, NY, USA</publisher-loc>
        <publisher-name>ACM</publisher-name>
        <fpage>1235</fpage>  
        <lpage>1244</lpage>  
        <pub-id pub-id-type="doi">10.1145/2783258.2783273</pub-id></nlm-citation>
      </ref>
      <ref id="ref12">
        <label>12</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>DH</given-names>
          </name>
          <name name-style="western">
            <surname>Liang</surname>
            <given-names>YC</given-names>
          </name>
          <name name-style="western">
            <surname>Xu</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Feng</surname>
            <given-names>XY</given-names>
          </name>
          <name name-style="western">
            <surname>Guan</surname>
            <given-names>RC</given-names>
          </name>
        </person-group>
        <article-title>A content-based recommender system for computer science publications</article-title>
        <source>Knowledge-Based Systems</source>  
        <year>2018</year>  
        <month>10</month>  
        <volume>157</volume>  
        <fpage>1</fpage>  
        <lpage>9</lpage>  
        <pub-id pub-id-type="doi">10.1016/j.knosys.2018.05.001</pub-id></nlm-citation>
      </ref>
      <ref id="ref13">
        <label>13</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Agarwal</surname>
            <given-names>V</given-names>
          </name>
          <name name-style="western">
            <surname>Zhang</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Zhu</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Fang</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Cheng</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Hong</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Shah</surname>
            <given-names>N</given-names>
          </name>
        </person-group>
        <article-title>Impact of Predicting Health Care Utilization Via Web Search Behavior: A Data-Driven Analysis</article-title>
        <source>J Med Internet Res</source>  
        <year>2016</year>  
        <month>09</month>  
        <day>21</day>  
        <volume>18</volume>  
        <issue>9</issue>  
        <fpage>e251</fpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmir.org/2016/9/e251/"/>
        </comment>  
        <pub-id pub-id-type="doi">10.2196/jmir.6240</pub-id>
        <pub-id pub-id-type="medline">27655225</pub-id>
        <pub-id pub-id-type="pii">v18i9e251</pub-id>
        <pub-id pub-id-type="pmcid">PMC5052461</pub-id></nlm-citation>
      </ref>
      <ref id="ref14">
        <label>14</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Blei</surname>
            <given-names>DM</given-names>
          </name>
          <name name-style="western">
            <surname>Ng</surname>
            <given-names>AY</given-names>
          </name>
          <name name-style="western">
            <surname>Jordan</surname>
            <given-names>MI</given-names>
          </name>
        </person-group>
        <article-title>Latent Dirichlet Allocation</article-title>
        <source>J Mach Learn Res</source>  
        <year>2003</year>  
        <month>1</month>  
        <volume>3</volume>  
        <fpage>993</fpage>  
        <lpage>1022</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmlr.org/papers/volume3/blei03a/blei03a.pdf"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref15">
        <label>15</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Altman</surname>
            <given-names>NS</given-names>
          </name>
        </person-group>
        <article-title>An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression</article-title>
        <source>The American Statistician</source>  
        <year>1992</year>  
        <month>08</month>  
        <volume>46</volume>  
        <issue>3</issue>  
        <fpage>175</fpage>  
        <lpage>185</lpage>  
        <pub-id pub-id-type="doi">10.1080/00031305.1992.10475879</pub-id></nlm-citation>
      </ref>
      <ref id="ref16">
        <label>16</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Yang</surname>
            <given-names>ZH</given-names>
          </name>
          <name name-style="western">
            <surname>Davison</surname>
            <given-names>BD</given-names>
          </name>
        </person-group>
        <article-title>Venue Recommendation: Submitting Your Paper with Style</article-title>
        <source>Proceedings of the 11th International Conference on Machine Learning and Applications</source>  
        <year>2012</year>  
        <conf-name>The 11th International Conference on Machine Learning and Applications</conf-name>
        <conf-date>2012 Dec 12-15</conf-date>
        <conf-loc>Boca Raton, FL</conf-loc>
        <fpage>681</fpage>  
        <lpage>686</lpage>  
        <pub-id pub-id-type="doi">10.1109/ICMLA.2012.127</pub-id></nlm-citation>
      </ref>
      <ref id="ref17">
        <label>17</label>
        <nlm-citation citation-type="web">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Küçüktunç</surname>
            <given-names>O</given-names>
          </name>
          <name name-style="western">
            <surname>Saule</surname>
            <given-names>E</given-names>
          </name>
          <name name-style="western">
            <surname>Kaya</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Çatalyürek</surname>
            <given-names>ÜV</given-names>
          </name>
        </person-group>
        <source>Arxiv.org</source>  
        <year>2012</year>  
        <month>05</month>  
        <day>05</day>  
        <access-date>2019-05-02</access-date>
        <comment>Recommendation on Academic Networks using Direction Aware Citation Analysis 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/pdf/1205.1143.pdf">https://arxiv.org/pdf/1205.1143.pdf</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="784Anxdv1"/></comment> </nlm-citation>
      </ref>
      <ref id="ref18">
        <label>18</label>
        <nlm-citation citation-type="book">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Luong</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Huynh</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Gauch</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Do</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Hoang</surname>
            <given-names>K</given-names>
          </name>
        </person-group>
        <person-group person-group-type="editor">
          <name name-style="western">
            <surname>Pan</surname>
            <given-names>JS</given-names>
          </name>
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>SM</given-names>
          </name>
          <name name-style="western">
            <surname>Nguyen</surname>
            <given-names>NT</given-names>
          </name>
        </person-group>
        <article-title>Publication Venue Recommendation Using Author Network's Publication History</article-title>
        <source>Intelligent Information and Database Systems. ACIIDS 2012.Lecture Notes in Computer Science</source>  
        <year>2012</year>  
        <publisher-loc>Berlin, Heidelberg</publisher-loc>
        <publisher-name>Springer</publisher-name>
        <fpage>426</fpage>  
        <lpage>435</lpage> </nlm-citation>
      </ref>
      <ref id="ref19">
        <label>19</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Beel</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Gipp</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Langer</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Breitinger</surname>
            <given-names>C</given-names>
          </name>
        </person-group>
        <article-title>Research-paper recommender systems: a literature survey</article-title>
        <source>Int J Digit Libr</source>  
        <year>2016</year>  
        <month>11</month>  
        <volume>17</volume>  
        <issue>4</issue>  
        <fpage>305</fpage>  
        <lpage>338</lpage>  
        <pub-id pub-id-type="doi">10.1007/s00799-015-0156-0</pub-id></nlm-citation>
      </ref>
      <ref id="ref20">
        <label>20</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Zhao</surname>
            <given-names>Z</given-names>
          </name>
          <name name-style="western">
            <surname>Huang</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Wang</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Lai</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Yu</surname>
            <given-names>P</given-names>
          </name>
        </person-group>
        <article-title>Low-Rank and Sparse Matrix Completion for Recommendation</article-title>
        <source>Proceedings of International Conference on Neural Information Processing</source>  
        <year>2017</year>  
        <conf-name>International Conference on Neural Information Processing</conf-name>
        <conf-date>Nov 14-18, 2017</conf-date>
        <conf-loc>Guangzhou, China</conf-loc>
        <fpage>3</fpage>  
        <lpage>13</lpage> </nlm-citation>
      </ref>
      <ref id="ref21">
        <label>21</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Luong</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Huynh</surname>
            <given-names>T</given-names>
          </name>
          <name name-style="western">
            <surname>Gauch</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Hoang</surname>
            <given-names>K</given-names>
          </name>
        </person-group>
        <article-title>Exploiting social networks for publication venue recommendations</article-title>
        <source>Proceedings of the International Conference on Knowledge Discovery and Information Retrieval</source>  
        <year>2012</year>  
        <conf-name>4th International Conference on Knowledge Discovery and Information Retrieval</conf-name>
        <conf-date>Oct 4-7, 2012</conf-date>
        <conf-loc>Barcelona, Spain</conf-loc>
        <fpage>239</fpage>  
        <lpage>245</lpage> </nlm-citation>
      </ref>
      <ref id="ref22">
        <label>22</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Salakhutdinov</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Mnih</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Hinton</surname>
            <given-names>G</given-names>
          </name>
        </person-group>
        <article-title>Restricted Boltzmann Machines for Collaborative Filtering</article-title>
        <source>Proceedings of the 24th International Conference on Machine Learning</source>  
        <year>2007</year>  
        <conf-name>The 24th International Conference on Machine Learning</conf-name>
        <conf-date>Jun 20-24, 2007</conf-date>
        <conf-loc>Corvalis, Oregon</conf-loc>
        <fpage>791</fpage>  
        <lpage>798</lpage>  
        <pub-id pub-id-type="doi">10.1145/1273496.1273596</pub-id></nlm-citation>
      </ref>
      <ref id="ref23">
        <label>23</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>McAuley</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Targett</surname>
            <given-names>C</given-names>
          </name>
          <name name-style="western">
            <surname>Shi</surname>
            <given-names>Q</given-names>
          </name>
          <name name-style="western">
            <surname>van den Hengel</surname>
            <given-names>A</given-names>
          </name>
        </person-group>
        <article-title>Image-Based Recommendations on StylesSubstitutes</article-title>
        <source>Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>  
        <year>2015</year>  
        <conf-name>The 38th International ACM SIGIR Conference on Research and Development in Information Retrieval</conf-name>
        <conf-date>Aug 09-13, 2015</conf-date>
        <conf-loc>Santiago, Chile</conf-loc>
        <fpage>43</fpage>  
        <lpage>52</lpage>  
        <pub-id pub-id-type="doi">10.1145/2766462.2767755</pub-id></nlm-citation>
      </ref>
      <ref id="ref24">
        <label>24</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>van den Oord</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Dieleman</surname>
            <given-names>S</given-names>
          </name>
          <name name-style="western">
            <surname>Schrauwen</surname>
            <given-names>B</given-names>
          </name>
        </person-group>
        <article-title>Deep content-based music recommendation</article-title>
        <source>Proceedings of the 27th Conference on Neural Information Processing Systems</source>  
        <year>2013</year>  
        <conf-name>the 27th Conference on Neural Information Processing Systems</conf-name>
        <conf-date>2013 Dec 05-10</conf-date>
        <conf-loc>Lake Tahoe, CA, USA</conf-loc>
        <fpage>2643</fpage>  
        <lpage>2651</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://papers.nips.cc/paper/5004-deep-content-based-music-recommendation.pdf"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref25">
        <label>25</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Collobert</surname>
            <given-names>R</given-names>
          </name>
          <name name-style="western">
            <surname>Weston</surname>
            <given-names>J</given-names>
          </name>
          <name name-style="western">
            <surname>Bottou</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Karlen</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Kavukcuoglu</surname>
            <given-names>K</given-names>
          </name>
          <name name-style="western">
            <surname>Kuksa</surname>
            <given-names>P</given-names>
          </name>
        </person-group>
        <article-title>Natural Language Processing (Almost) from Scratch</article-title>
        <source>Journal of Machine Learning Research</source>  
        <year>2011</year>  
        <month>08</month>  
        <volume>12</volume>  
        <fpage>2493</fpage>  
        <lpage>2537</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://www.jmlr.org/papers/volume12/collobert11a/collobert11a.pdf"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref26">
        <label>26</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Bengio</surname>
            <given-names>Y</given-names>
          </name>
          <name name-style="western">
            <surname>Courville</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Vincent</surname>
            <given-names>P</given-names>
          </name>
        </person-group>
        <article-title>Representation learning: a review and new perspectives</article-title>
        <source>IEEE Trans Pattern Anal Mach Intell</source>  
        <year>2013</year>  
        <month>08</month>  
        <volume>35</volume>  
        <issue>8</issue>  
        <fpage>1798</fpage>  
        <lpage>1828</lpage>  
        <pub-id pub-id-type="doi">10.1109/TPAMI.2013.50</pub-id>
        <pub-id pub-id-type="medline">23787338</pub-id></nlm-citation>
      </ref>
      <ref id="ref27">
        <label>27</label>
        <nlm-citation citation-type="web">
        <source>NCBI</source>  
        <access-date>2019-05-02</access-date>
        <comment>FTP Service 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://www.ncbi.nlm.nih.gov/pmc/tools/ftp/">https://www.ncbi.nlm.nih.gov/pmc/tools/ftp/</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="784GfhZXX"/></comment> </nlm-citation>
      </ref>
      <ref id="ref28">
        <label>28</label>
        <nlm-citation citation-type="web">
        <source>Natural Language Toolkit</source>  
        <access-date>2018-11-26</access-date>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://www.nltk.org/">https://www.nltk.org/</ext-link>
          <ext-link ext-link-type="webcite" xlink:href="74E8QTItd"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref29">
        <label>29</label>
        <nlm-citation citation-type="web">
        <source>EVEX</source>  
        <access-date>2018-11-26</access-date>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://evexdb.org/pmresources/vec-space-models/">http://evexdb.org/pmresources/vec-space-models/</ext-link>
          <ext-link ext-link-type="webcite" xlink:href="74E8pH7A4"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref30">
        <label>30</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Hu</surname>
            <given-names>B</given-names>
          </name>
          <name name-style="western">
            <surname>Lu</surname>
            <given-names>Z</given-names>
          </name>
          <name name-style="western">
            <surname>Li</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Chen</surname>
            <given-names>Q</given-names>
          </name>
        </person-group>
        <article-title>Convolutional Neural Network Architectures for Matching Natural Language Sentences</article-title>
        <source>Proceeding of 27th Conference on Neural Information Processing Systems</source>  
        <year>2014</year>  
        <conf-name>27th Conference on Neural Information Processing Systems</conf-name>
        <conf-date>Dec 8-13, 2014</conf-date>
        <conf-loc>Montreal, Canada</conf-loc>
        <fpage>2042</fpage>  
        <lpage>2050</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="http://papers.nips.cc/paper/5550-convolutional-neural-network-architectures-for-matching-natural-language-sentences.pdf"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref31">
        <label>31</label>
        <nlm-citation citation-type="web">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Kingma</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Ba</surname>
            <given-names>J</given-names>
          </name>
        </person-group>
        <source>ArXiv.org</source>  
        <year>2014</year>  
        <month>12</month>  
        <day>22</day>  
        <access-date>2019-05-02</access-date>
        <comment>Adam: A Method for Stochastic Optimization 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://arxiv.org/abs/1412.6980">https://arxiv.org/abs/1412.6980</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="7846ZaGfU"/></comment> </nlm-citation>
      </ref>
      <ref id="ref32">
        <label>32</label>
        <nlm-citation citation-type="web">
        <source>GitHub</source>  
        <year>2018</year>  
        <access-date>2018-11-27</access-date>
        <comment>FastText 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://github.com/facebookresearch/fastText">https://github.com/facebookresearch/fastText</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="74E8vJ40r"/></comment> </nlm-citation>
      </ref>
      <ref id="ref33">
        <label>33</label>
        <nlm-citation citation-type="confproc">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Conneau</surname>
            <given-names>A</given-names>
          </name>
          <name name-style="western">
            <surname>Kiela</surname>
            <given-names>D</given-names>
          </name>
          <name name-style="western">
            <surname>Schwenk</surname>
            <given-names>H</given-names>
          </name>
          <name name-style="western">
            <surname>Barrault</surname>
            <given-names>L</given-names>
          </name>
          <name name-style="western">
            <surname>Bordes</surname>
            <given-names>A</given-names>
          </name>
        </person-group>
        <article-title>Supervised Learning of Universal Sentence Representations from Natural Language Inference Data</article-title>
        <source>Proceedings of 2017 Conference on Empirical Methods in Natural Language Processing</source>  
        <year>2017</year>  
        <conf-name>EMNLP 2017: Conference on Empirical Methods in Natural Language Processing</conf-name>
        <conf-date>Sep 7-11, 2017</conf-date>
        <conf-loc>Copenhagen, Denmark</conf-loc>
        <fpage>670</fpage>  
        <lpage>680</lpage>  
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://www.aclweb.org/anthology/D17-1070"/>
        </comment> </nlm-citation>
      </ref>
      <ref id="ref34">
        <label>34</label>
        <nlm-citation citation-type="journal">
        <person-group person-group-type="author">
          <name name-style="western">
            <surname>Sokolova</surname>
            <given-names>M</given-names>
          </name>
          <name name-style="western">
            <surname>Lapalme</surname>
            <given-names>G</given-names>
          </name>
        </person-group>
        <article-title>A systematic analysis of performance measures for classification tasks</article-title>
        <source>Information Processing &amp; Management</source>  
        <year>2009</year>  
        <month>7</month>  
        <volume>45</volume>  
        <issue>4</issue>  
        <fpage>427</fpage>  
        <lpage>437</lpage>  
        <pub-id pub-id-type="doi">10.1016/j.ipm.2009.03.002</pub-id></nlm-citation>
      </ref>
      <ref id="ref35">
        <label>35</label>
        <nlm-citation citation-type="web">
        <source>Elsevier</source>  
        <access-date>2018-11-26</access-date>
        <comment>Journal Finder 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://journalfinder.elsevier.com/">https://journalfinder.elsevier.com/</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="74E93AryP"/></comment> </nlm-citation>
      </ref>
      <ref id="ref36">
        <label>36</label>
        <nlm-citation citation-type="web">
        <source>Springer Nature</source>  
        <access-date>2018-11-26</access-date>
        <comment>Journal Suggester 
        <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://journalsuggester.springer.com/">https://journalsuggester.springer.com/</ext-link>
        <ext-link ext-link-type="webcite" xlink:href="74E9DA7tZ"/></comment> </nlm-citation>
      </ref>
      <ref id="ref37">
        <label>37</label>
        <nlm-citation citation-type="web">
        <source>Pubmender: Deep Learning Based Recommender System for Biomedical Publication Venue</source>  
        <access-date>2019-05-07</access-date>
        <comment>
          <ext-link xmlns:xlink="http://www.w3.org/1999/xlink" ext-link-type="uri" xlink:type="simple" xlink:href="https://www.keaml.cn:8081/">https://www.keaml.cn:8081/</ext-link>
          <ext-link ext-link-type="webcite" xlink:href="78CMVBjkL"/>
        </comment> </nlm-citation>
      </ref>
    </ref-list>
  </back>
</article>
