<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.3 20210610//EN" "JATS-journalpublishing1-3.dtd">
<article article-type="research-article" dtd-version="1.3" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xml:lang="ru"><front><journal-meta><journal-id journal-id-type="publisher-id">vestnikmephi</journal-id><journal-title-group><journal-title xml:lang="ru">Вестник НИЯУ МИФИ</journal-title><trans-title-group xml:lang="en"><trans-title>Vestnik natsional'nogo issledovatel'skogo yadernogo universiteta "MIFI"</trans-title></trans-title-group></journal-title-group><issn pub-type="ppub">2304-487X</issn><publisher><publisher-name>National Research Nuclear University "MEPhI"</publisher-name></publisher></journal-meta><article-meta><article-id pub-id-type="doi">10.1134/S2304487X19060129</article-id><article-id custom-type="elpub" pub-id-type="custom">vestnikmephi-68</article-id><article-categories><subj-group subj-group-type="heading"><subject>Research Article</subject></subj-group><subj-group subj-group-type="section-heading" xml:lang="ru"><subject>ПРИКЛАДНАЯ МАТЕМАТИКА И ИНФОРМАТИКА</subject></subj-group><subj-group subj-group-type="section-heading" xml:lang="en"><subject>APPLIED MATHEMATICS AND COMPUTER SCIENCE</subject></subj-group></article-categories><title-group><article-title>Генеративно-дискриминативная нейросетевая модель для задачи авторского профилирования</article-title><trans-title-group xml:lang="en"><trans-title>Generative-Discriminative Neural Network Model for the Task of Author Profiling</trans-title></trans-title-group></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name-alternatives><name name-style="eastern" xml:lang="ru"><surname>Сбоев</surname><given-names>А. Г.</given-names></name><name name-style="western" xml:lang="en"><surname>Sboev</surname><given-names>A. G.</given-names></name></name-alternatives><bio xml:lang="ru"><p>123098</p><p>115409</p><p>Москва</p></bio><bio xml:lang="en"><p>123098</p><p>115409</p><p>Moscow</p></bio><email xlink:type="simple">sag111@mail.ru</email><xref ref-type="aff" rid="aff-1"/></contrib><contrib contrib-type="author" corresp="yes"><name-alternatives><name name-style="eastern" xml:lang="ru"><surname>Рыбка</surname><given-names>Р. Б.</given-names></name><name name-style="western" xml:lang="en"><surname>Rybka</surname><given-names>R. B.</given-names></name></name-alternatives><bio xml:lang="ru"><p>123098</p><p>Москва</p></bio><bio xml:lang="en"><p>123098</p><p>Moscow</p></bio><xref ref-type="aff" rid="aff-2"/></contrib><contrib contrib-type="author" corresp="yes"><name-alternatives><name name-style="eastern" xml:lang="ru"><surname>Грязнов</surname><given-names>А. В.</given-names></name><name name-style="western" xml:lang="en"><surname>Gryaznov</surname><given-names>A. V.</given-names></name></name-alternatives><bio xml:lang="ru"><p>123098</p><p>Москва</p></bio><bio xml:lang="en"><p>123098</p><p>Moscow</p></bio><xref ref-type="aff" rid="aff-2"/></contrib><contrib contrib-type="author" corresp="yes"><name-alternatives><name name-style="eastern" xml:lang="ru"><surname>Молошников</surname><given-names>И. А.</given-names></name><name name-style="western" xml:lang="en"><surname>Moloshnikov</surname><given-names>I. A.</given-names></name></name-alternatives><bio xml:lang="ru"><p>123098</p><p>Москва</p></bio><bio xml:lang="en"><p>123098</p><p>Moscow</p></bio><xref ref-type="aff" rid="aff-2"/></contrib></contrib-group><aff-alternatives id="aff-1"><aff xml:lang="ru">Национальный исследовательский центр “Курчатовский институт”; Национальный исследовательский ядерный университет “МИФИ”<country>Россия</country></aff><aff xml:lang="en">National Research Center “Kurchatov Institute”; National Research Nuclear University “MEPhI” (Moscow Engineering Physics Institute)<country>Russian Federation</country></aff></aff-alternatives><aff-alternatives id="aff-2"><aff xml:lang="ru">Национальный исследовательский центр “Курчатовский институт”<country>Россия</country></aff><aff xml:lang="en">National Research Center “Kurchatov Institute”<country>Russian Federation</country></aff></aff-alternatives><pub-date pub-type="collection"><year>2020</year></pub-date><pub-date pub-type="epub"><day>12</day><month>02</month><year>2023</year></pub-date><volume>9</volume><issue>1</issue><fpage>50</fpage><lpage>57</lpage><permissions><copyright-statement>Copyright &amp;#x00A9; Сбоев А.Г., Рыбка Р.Б., Грязнов А.В., Молошников И.А., 2023</copyright-statement><copyright-year>2023</copyright-year><copyright-holder xml:lang="ru">Сбоев А.Г., Рыбка Р.Б., Грязнов А.В., Молошников И.А.</copyright-holder><copyright-holder xml:lang="en">Sboev A.G., Rybka R.B., Gryaznov A.V., Moloshnikov I.A.</copyright-holder><license license-type="creative-commons-attribution" xlink:href="https://creativecommons.org/licenses/by/4.0/" xlink:type="simple"><license-p>This work is licensed under a Creative Commons Attribution 4.0 License.</license-p></license></permissions><self-uri xlink:href="https://vestnikmephi.elpub.ru/jour/article/view/68">https://vestnikmephi.elpub.ru/jour/article/view/68</self-uri><abstract/><trans-abstract xml:lang="en"><p>   The paper considers the generative-discriminative model (GAN) as applied to the task of analyzing text data, in particular, determining the gender of the author of a Russian-language text. The approach developed belongs to semi-supervised learning algorithms, when both labeled and unlabeled samples are involved in the model fitting process. The GAN model is implemented as a deep neural network consisting of fully connected, recurrent and convolution layers. The basis of the generative part of the GAN model is a variational auto-encoder, which encodes the input sample into the space of hidden variables and then decodes the latter into the original representation. When decoding, the class label of the input example is used, known in the case of the labeled set or predicted by the classifier for unlabeled samples. The input for the model is a sequence of words, each encoded by a vector of the principal components of its morphological features. To provide the function of recovering texts with more than 50 words, the principles of the work of language models are used. The discriminant part is configured to determine whether a given sample was generated by an auto-encoder or taken from the original set. Quality assessment of the GAN model was carried out on a set of texts from LiveJournal blogs. It is shown that the use of the generative-discriminative model allows to improve the quality of classification by 2 % in the F1 metric, while reducing the standard deviation by 2–3 times when learning on a small number of labeled examples. Various modes of training and variations in the topology of the GAN model are investigated, and the most effective modes of operation of models of this type for the task of classifying texts are demonstrated.</p></trans-abstract><kwd-group xml:lang="ru"><kwd>машинное обучение</kwd><kwd>искусственные нейронные сети</kwd><kwd>обработка естественного языка</kwd><kwd>автоматизированный анализ текстов</kwd><kwd>генеративно-дискриминативные нейронные сети</kwd><kwd>авторское профилирование</kwd><kwd>определение пола автора текста</kwd></kwd-group><kwd-group xml:lang="en"><kwd>machine learning</kwd><kwd>artificial neural networks</kwd><kwd>natural language processing</kwd><kwd>automated text analysis</kwd><kwd>generative-discriminative neural networks</kwd><kwd>author profiling</kwd><kwd>author gender identification</kwd></kwd-group></article-meta></front><back><ref-list><title>References</title><ref id="cit1"><label>1</label><citation-alternatives><mixed-citation xml:lang="ru">Guo J. et al. Long text generation via adversarial training with leaked information // Thirty-Second AAAI Conference on Artificial Intelligence. 2018.</mixed-citation><mixed-citation xml:lang="en">Guo J. et al. Long text generation via adversarial training with leaked information // Thirty-Second AAAI Conference on Artificial Intelligence. 2018.</mixed-citation></citation-alternatives></ref><ref id="cit2"><label>2</label><citation-alternatives><mixed-citation xml:lang="ru">Nie W., Narodytska N., Patel A. RelGAN: Relational generative adversarial networks for text generation. 2018.</mixed-citation><mixed-citation xml:lang="en">Nie W., Narodytska N., Patel A. RelGAN: Relational generative adversarial networks for text generation. 2018.</mixed-citation></citation-alternatives></ref><ref id="cit3"><label>3</label><citation-alternatives><mixed-citation xml:lang="ru">Radford A. et al. Language models are unsupervised multitask learners // OpenAI Blog. 2019. Vol. 1. № 8.</mixed-citation><mixed-citation xml:lang="en">Radford A. et al. Language models are unsupervised multitask learners // OpenAI Blog. 2019. Vol. 1. № 8.</mixed-citation></citation-alternatives></ref><ref id="cit4"><label>4</label><citation-alternatives><mixed-citation xml:lang="ru">Bao J. et al. CVAE-GAN: fine-grained image generation through asymmetric training // Proceedings of the IEEE International Conference on Computer Vision. 2017. С. 2745–2754.</mixed-citation><mixed-citation xml:lang="en">Bao J. et al. CVAE-GAN: fine-grained image generation through asymmetric training // Proceedings of the IEEE International Conference on Computer Vision. 2017. С. 2745–2754.</mixed-citation></citation-alternatives></ref><ref id="cit5"><label>5</label><citation-alternatives><mixed-citation xml:lang="ru">Xu W. et al. Variational autoencoder for semi-supervised text classification // Thirty-First AAAI Conference on Artificial Intelligence. 2017.</mixed-citation><mixed-citation xml:lang="en">Xu W. et al. Variational autoencoder for semi-supervised text classification // Thirty-First AAAI Conference on Artificial Intelligence. 2017.</mixed-citation></citation-alternatives></ref><ref id="cit6"><label>6</label><citation-alternatives><mixed-citation xml:lang="ru">Semeniuta S., Severyn A., Barth E. A hybrid convolutional variational autoencoder for text generation // arXiv preprint arXiv:1702.02390. 2017.</mixed-citation><mixed-citation xml:lang="en">Semeniuta S., Severyn A., Barth E. A hybrid convolutional variational autoencoder for text generation // arXiv preprint arXiv:1702.02390. 2017.</mixed-citation></citation-alternatives></ref><ref id="cit7"><label>7</label><citation-alternatives><mixed-citation xml:lang="ru">Shen D. et al. Hierarchically-Structured Variational Autoencoders for Long Text Generation. 2018.</mixed-citation><mixed-citation xml:lang="en">Shen D. et al. Hierarchically-Structured Variational Autoencoders for Long Text Generation. 2018.</mixed-citation></citation-alternatives></ref><ref id="cit8"><label>8</label><citation-alternatives><mixed-citation xml:lang="ru">Chung J. et al. Empirical evaluation of gated recurrent neural networks on sequence modeling // arXiv preprint arXiv:1412.3555. 2014.</mixed-citation><mixed-citation xml:lang="en">Chung J. et al. Empirical evaluation of gated recurrent neural networks on sequence modeling // arXiv preprint arXiv:1412.3555. 2014.</mixed-citation></citation-alternatives></ref><ref id="cit9"><label>9</label><citation-alternatives><mixed-citation xml:lang="ru">Chollet F. et al. Keras. 2015.</mixed-citation><mixed-citation xml:lang="en">Chollet F. et al. Keras. 2015.</mixed-citation></citation-alternatives></ref><ref id="cit10"><label>10</label><citation-alternatives><mixed-citation xml:lang="ru">Abadi M. et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems // arXiv preprint arXiv:1603.04467. 2016.</mixed-citation><mixed-citation xml:lang="en">Abadi M. et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems // arXiv preprint arXiv:1603.04467. 2016.</mixed-citation></citation-alternatives></ref><ref id="cit11"><label>11</label><citation-alternatives><mixed-citation xml:lang="ru">Arjovsky M., Shah A., Bengio Y. Unitary evolution recurrent neural networks // International Conference on Machine Learning. 2016. P. 1120–1128.</mixed-citation><mixed-citation xml:lang="en">Arjovsky M., Shah A., Bengio Y. Unitary evolution recurrent neural networks // International Conference on Machine Learning. 2016. P. 1120–1128.</mixed-citation></citation-alternatives></ref><ref id="cit12"><label>12</label><citation-alternatives><mixed-citation xml:lang="ru">Kingma D. P., Ba J. Adam: A method for stochastic optimization // arXiv preprint arXiv:1412.6980. 2014.</mixed-citation><mixed-citation xml:lang="en">Kingma D. P., Ba J. Adam: A method for stochastic optimization // arXiv preprint arXiv:1412.6980. 2014.</mixed-citation></citation-alternatives></ref><ref id="cit13"><label>13</label><citation-alternatives><mixed-citation xml:lang="ru">Litvinova T. A., Sboev A. G., Panicheva P. V. Profiling the Age of Russian Bloggers // Conference on Artificial Intelligence and Natural Language. Springer, Cham, 2018. P. 167–177.</mixed-citation><mixed-citation xml:lang="en">Litvinova T. A., Sboev A. G., Panicheva P. V. Profiling the Age of Russian Bloggers // Conference on Artificial Intelligence and Natural Language. Springer, Cham, 2018. P. 167–177.</mixed-citation></citation-alternatives></ref><ref id="cit14"><label>14</label><citation-alternatives><mixed-citation xml:lang="ru">Straka M., Hajic J., Straková J. UDPipe: trainable pipeline for processing CoNLL-U files performing tokenization, morphological analysis, pos tagging and parsing // Proceedings of the tenth international conference on language resources and evaluation (LREC 2016). 2016. P. 4290–4297.</mixed-citation><mixed-citation xml:lang="en">Straka M., Hajic J., Straková J. UDPipe: trainable pipeline for processing CoNLL-U files performing tokenization, morphological analysis, pos tagging and parsing // Proceedings of the tenth international conference on language resources and evaluation (LREC 2016). 2016. P. 4290–4297.</mixed-citation></citation-alternatives></ref><ref id="cit15"><label>15</label><citation-alternatives><mixed-citation xml:lang="ru">Tipping M. E., Bishop C. M. Probabilistic principal component analysis // Journal of the Royal Statistical Society: Series B (Statistical Methodology).</mixed-citation><mixed-citation xml:lang="en">Tipping M. E., Bishop C. M. Probabilistic principal component analysis // Journal of the Royal Statistical Society: Series B (Statistical Methodology).</mixed-citation></citation-alternatives></ref></ref-list><fn-group><fn fn-type="conflict"><p>The authors declare that there are no conflicts of interest present.</p></fn></fn-group></back></article>
