<!DOCTYPE article
PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.4 20190208//EN"
       "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" article-type="research-article" dtd-version="1.4" xml:lang="en">
 <front>
  <journal-meta>
   <journal-id journal-id-type="publisher-id">Journal of Technical Research</journal-id>
   <journal-title-group>
    <journal-title xml:lang="en">Journal of Technical Research</journal-title>
    <trans-title-group xml:lang="ru">
     <trans-title>Журнал технических исследований</trans-title>
    </trans-title-group>
   </journal-title-group>
   <issn publication-format="print">2500-3313</issn>
  </journal-meta>
  <article-meta>
   <article-id pub-id-type="publisher-id">106302</article-id>
   <article-categories>
    <subj-group subj-group-type="toc-heading" xml:lang="ru">
     <subject>Информационные технологии и телекоммуникации</subject>
    </subj-group>
    <subj-group subj-group-type="toc-heading" xml:lang="en">
     <subject>Information technology and telecommunication</subject>
    </subj-group>
    <subj-group>
     <subject>Информационные технологии и телекоммуникации</subject>
    </subj-group>
   </article-categories>
   <title-group>
    <article-title xml:lang="en">Application of machine learning methods  and mathematical models  for big data analysis</article-title>
    <trans-title-group xml:lang="ru">
     <trans-title>Применение методов машинного обучения и математических моделей для анализа больших данных</trans-title>
    </trans-title-group>
   </title-group>
   <contrib-group content-type="authors">
    <contrib contrib-type="author">
     <name-alternatives>
      <name xml:lang="ru">
       <surname>Исаева</surname>
       <given-names>А. Т.</given-names>
      </name>
      <name xml:lang="en">
       <surname>Isaeva</surname>
       <given-names>Aida Taalaevna</given-names>
      </name>
     </name-alternatives>
     <email>akeldibekova@oshsu.kg</email>
     <xref ref-type="aff" rid="aff-1"/>
    </contrib>
    <contrib contrib-type="author">
     <contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-6444-0468</contrib-id>
     <name-alternatives>
      <name xml:lang="ru">
       <surname>Келдибекова</surname>
       <given-names>А. О.</given-names>
      </name>
      <name xml:lang="en">
       <surname>Keldibekova</surname>
       <given-names>Aida Oskonovna</given-names>
      </name>
     </name-alternatives>
     <email>aidaoskk@gmail.com</email>
     <bio xml:lang="ru">
      <p>доктор педагогических наук;</p>
     </bio>
     <bio xml:lang="en">
      <p>doctor of pedagogical sciences;</p>
     </bio>
     <xref ref-type="aff" rid="aff-2"/>
    </contrib>
   </contrib-group>
   <aff-alternatives id="aff-1">
    <aff>
     <institution xml:lang="ru">Ошский государственный университет</institution>
     <city>Osh</city>
     <country>KG</country>
    </aff>
    <aff>
     <institution xml:lang="en">Osh State University</institution>
     <city>Osh</city>
     <country>KG</country>
    </aff>
   </aff-alternatives>
   <aff-alternatives id="aff-2">
    <aff>
     <institution xml:lang="ru">Ошский государственный университет</institution>
     <city>Ош</city>
     <country>Киргизия</country>
    </aff>
    <aff>
     <institution xml:lang="en">Osh state university</institution>
     <city>Osh</city>
     <country>Kyrgyzstan</country>
    </aff>
   </aff-alternatives>
   <pub-date publication-format="print" date-type="pub" iso-8601-date="2025-12-30T00:00:00+03:00">
    <day>30</day>
    <month>12</month>
    <year>2025</year>
   </pub-date>
   <pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-12-30T00:00:00+03:00">
    <day>30</day>
    <month>12</month>
    <year>2025</year>
   </pub-date>
   <volume>11</volume>
   <issue>4</issue>
   <fpage>23</fpage>
   <lpage>27</lpage>
   <history>
    <date date-type="received" iso-8601-date="2025-11-07T00:00:00+03:00">
     <day>07</day>
     <month>11</month>
     <year>2025</year>
    </date>
    <date date-type="accepted" iso-8601-date="2025-11-14T00:00:00+03:00">
     <day>14</day>
     <month>11</month>
     <year>2025</year>
    </date>
   </history>
   <self-uri xlink:href="https://naukaru.ru/en/nauka/article/106302/view">https://naukaru.ru/en/nauka/article/106302/view</self-uri>
   <abstract xml:lang="ru">
    <p>В условиях стремительного роста объемов информации проблема анализа больших данных &#13;
(Big Data) приобретает особую актуальность. В данной статье исследуется симбиоз методов машинного обучения (МО) и фундаментальных математических моделей как основа для эффективного извлечения знаний из больших массивов информации. Цель работы — разработка и сравнительная оценка комплекса методов МО, подкрепленных математическим аппаратом, для задач классификации и кластеризации. На основе эксперимента с использованием набора данных UCI Machine Learning Repository проведен сравнительный анализ алгоритмов, включая логистическую регрессию, метод опорных векторов (SVM), случайный лес и многослойный перцептрон. Результаты показывают, что нейронные сети (Accuracy: 0.92, F1-мера: 0.89) и ансамблевые методы демонстрируют превосходство над классическими алгоритмами при работе с разнородными данными. Подчеркивается, что математические модели из областей оптимизации, линейной алгебры и теории вероятностей являются неотъемлемым фундаментом, обеспечивающим корректность и эффективность алгоритмов МО. Делается вывод &#13;
о целесообразности комплексного подхода, объединяющего вычислительную мощь МО и строгость математических моделей.</p>
   </abstract>
   <trans-abstract xml:lang="en">
    <p>In the context of the rapid growth of information volumes, the problem of big data analysis is becoming particularly relevant. This article investigates the symbiosis of machine learning (ML) methods and fundamental mathematical models as a basis for effective knowledge extraction from large datasets. aree aim of the work is the development and comparative evaluation of a set of ML methods supported by a mathematical apparatus for classification and clustering tasks. Based on an experiment using a dataset from the UCI Machine Learning Repository, a comparative analysis of algorithms was conducted, including Logistic Regression, Support Vector Machine (SVM), Random Forest, and Multilayer Perceptron. The results show that neural networks (Accuracy: 0.92, F1-score: 0.89) and ensemble methods outperform classical algorithms when working with heterogeneous data. It is emphasized that mathematical models from the fields of optimization, linear algebra, and probability theory are an integral foundation that ensures the correctness and efficiency of ML algorithms. The conclusion is made about the feasibility of an integrated approach combining the computational power of ML and the rigor of mathematical models.</p>
   </trans-abstract>
   <kwd-group xml:lang="ru">
    <kwd>большие данные</kwd>
    <kwd>машинное обучение</kwd>
    <kwd>математические модели</kwd>
    <kwd>классификация</kwd>
    <kwd>кластеризация</kwd>
    <kwd>нейронные сети</kwd>
    <kwd>оптимизация</kwd>
   </kwd-group>
   <kwd-group xml:lang="en">
    <kwd>big data</kwd>
    <kwd>machine learning</kwd>
    <kwd>mathematical models</kwd>
    <kwd>classification</kwd>
    <kwd>clustering</kwd>
    <kwd>neural networks</kwd>
    <kwd>optimization</kwd>
   </kwd-group>
  </article-meta>
 </front>
 <body>
  <p></p>
 </body>
 <back>
  <ref-list>
   <ref id="B1">
    <label>1.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Chen, M., Mao, S., &amp; Liu, Y. Big Data: A Survey. Mobile Networks and Applications. 2019. №19(2). Pp. 171–209.</mixed-citation>
     <mixed-citation xml:lang="en">Chen, M., Mao, S., &amp; Liu, Y. Big Data: A Survey. Mobile Networks and Applications. 2019. №19(2). Pp. 171–209.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B2">
    <label>2.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Deisenroth, M. P., Faisal, A. A. &amp; Ong, C. S. Mathematics for Machine Learning. Cambridge University Press. 2020. 398 p.</mixed-citation>
     <mixed-citation xml:lang="en">Deisenroth, M. P., Faisal, A. A. &amp; Ong, C. S. Mathematics for Machine Learning. Cambridge University Press. 2020. 398 p.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B3">
    <label>3.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Bottou, L., Curtis, F. E., &amp; Nocedal, J. Optimization Methods for Large-Scale Machine Learning. SIAM Review. 2018. № 60(2). Pp. 223–311.</mixed-citation>
     <mixed-citation xml:lang="en">Bottou, L., Curtis, F. E., &amp; Nocedal, J. Optimization Methods for Large-Scale Machine Learning. SIAM Review. 2018. № 60(2). Pp. 223–311.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B4">
    <label>4.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Исаков, Р. Ж., Абдыкалыков, А. А. Возможности применения искусственного интеллекта в диагностике заболеваний в Кыргызстане. Вестник Кыргызско-Российского Славянского университета. 2022. № 22(5). С. 124–130.</mixed-citation>
     <mixed-citation xml:lang="en">Isakov, R. Zh., Abdykalykov, A. A. Possibilities of applying artificial intelligence in disease diagnostics in Kyrgyzstan. Bulletin of the Kyrgyz-Russian Slavic University. 2022. No. 22(5). Pp. 124–130.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B5">
    <label>5.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Murphy, K. P. Probabilistic Machine Learning: An Introduction. MIT Press. 2022. 864 p.</mixed-citation>
     <mixed-citation xml:lang="en">Murphy, K. P. Probabilistic Machine Learning: An Introduction. MIT Press. 2022. 864 p.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B6">
    <label>6.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., &amp; Gulin, A.  CatBoost: unbiased boosting with categorical features. Advances in Neural Information Processing Systems. 2018. P. 31.</mixed-citation>
     <mixed-citation xml:lang="en">Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., &amp; Gulin, A.  CatBoost: unbiased boosting with categorical features. Advances in Neural Information Processing Systems. 2018. P. 31.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B7">
    <label>7.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Goodfellow I., Bengio, Y. &amp; Courville, A. Deep Learning. MIT Press. 2016. 800 p.</mixed-citation>
     <mixed-citation xml:lang="en">Goodfellow I., Bengio, Y. &amp; Courville, A. Deep Learning. MIT Press. 2016. 800 p.</mixed-citation>
    </citation-alternatives>
   </ref>
   <ref id="B8">
    <label>8.</label>
    <citation-alternatives>
     <mixed-citation xml:lang="ru">Dua, D. and Graff, C. UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science. 2019. [Электронный ресурс]. URL: http://archive.ics.uci.edu/ml</mixed-citation>
     <mixed-citation xml:lang="en">Dua, D. and Graff, C. UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science. 2019. [Electronic resource]. URL: http://archive.ics.uci.edu/ml</mixed-citation>
    </citation-alternatives>
   </ref>
  </ref-list>
 </back>
</article>
