<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.4 20241031//EN"
        "https://jats.nlm.nih.gov/publishing/1.4/JATS-journalpublishing1-4.dtd">
<article  article-type="research-article"        dtd-version="1.4">
            <front>

                <journal-meta>
                                    <journal-id></journal-id>
            <journal-title-group>
                                                                                    <journal-title>Balkan Journal of Electrical and Computer Engineering</journal-title>
            </journal-title-group>
                            <issn pub-type="ppub">2147-284X</issn>
                                        <issn pub-type="epub">2147-284X</issn>
                                                                                            <publisher>
                    <publisher-name>Balkan Yayın</publisher-name>
                </publisher>
                    </journal-meta>
                <article-meta>
                                        <article-id pub-id-type="doi">10.17694/bajece.1114868</article-id>
                                                                <article-categories>
                                            <subj-group  xml:lang="en">
                                                            <subject>Computer Software</subject>
                                                            <subject>Software Architecture</subject>
                                                    </subj-group>
                                            <subj-group  xml:lang="tr">
                                                            <subject>Bilgisayar Yazılımı</subject>
                                                            <subject>Yazılım Mimarisi</subject>
                                                    </subj-group>
                                    </article-categories>
                                                                                                                                                        <title-group>
                                                                                                                        <article-title>An Approach for DC Motor Speed Control with Off-Policy Reinforcement Learning Method</article-title>
                                                                                                                                        </title-group>
            
                                                    <contrib-group content-type="authors">
                                                                        <contrib contrib-type="author">
                                                                    <contrib-id contrib-id-type="orcid">
                                        https://orcid.org/0000-0001-9815-7724</contrib-id>
                                                                <name>
                                    <surname>Tüfenkçi</surname>
                                    <given-names>Sevilay</given-names>
                                </name>
                                                                    <aff>MALATYA TURGUT ÖZAL ÜNİVERSİTESİ</aff>
                                                            </contrib>
                                                    <contrib contrib-type="author">
                                                                    <contrib-id contrib-id-type="orcid">
                                        https://orcid.org/0000-0003-2651-5005</contrib-id>
                                                                <name>
                                    <surname>Kavuran</surname>
                                    <given-names>Gürkan</given-names>
                                </name>
                                                                    <aff>MALATYA TURGUT ÖZAL ÜNİVERSİTESİ</aff>
                                                            </contrib>
                                                    <contrib contrib-type="author">
                                                                    <contrib-id contrib-id-type="orcid">
                                        https://orcid.org/0000-0002-6106-2374</contrib-id>
                                                                <name>
                                    <surname>Yeroğlu</surname>
                                    <given-names>Celaleddin</given-names>
                                </name>
                                                                    <aff>İNÖNÜ ÜNİVERSİTESİ</aff>
                                                            </contrib>
                                                                                </contrib-group>
                        
                                        <pub-date pub-type="pub" iso-8601-date="20230604">
                    <day>06</day>
                    <month>04</month>
                    <year>2023</year>
                </pub-date>
                                        <volume>11</volume>
                                        <issue>2</issue>
                                        <fpage>184</fpage>
                                        <lpage>189</lpage>
                        
                        <history>
                                    <date date-type="received" iso-8601-date="20220510">
                        <day>05</day>
                        <month>10</month>
                        <year>2022</year>
                    </date>
                                                    <date date-type="accepted" iso-8601-date="20230530">
                        <day>05</day>
                        <month>30</month>
                        <year>2023</year>
                    </date>
                            </history>
                                        <permissions>
                    <copyright-statement>Copyright © 2013, Balkan Journal of Electrical and Computer Engineering</copyright-statement>
                    <copyright-year>2013</copyright-year>
                    <copyright-holder>Balkan Journal of Electrical and Computer Engineering</copyright-holder>
                </permissions>
            
                                                                                                <abstract><p>In the literature, interest in automatic control systems that do not require human intervention and perform at the desired level increases day by day. In this study, a Twin Delay Deep Deterministic Policy Gradient (TD3), a reinforcement learning algorithm, automatically controls a DC motor system. A reinforcement learning method is an approach that learns what should be done to reach the goal and observes the results that come out with the interaction of both itself and the environment. The proposed method aims to adjust the voltage value applied to the input of the DC motor in order to reach output with single input and single output structure to the desired speed.</p></abstract>
                                                                                    
            
                                                            <kwd-group>
                                                    <kwd>Deep reinforcement learning</kwd>
                                                    <kwd>  DC motor</kwd>
                                                    <kwd>  PI controller</kwd>
                                                    <kwd>  Twin-delayed deep deterministic policy gradient</kwd>
                                            </kwd-group>
                                                        
                                                                                                                                                    </article-meta>
    </front>
    <back>
                            <ref-list>
                                    <ref id="ref1">
                        <label>1</label>
                        <mixed-citation publication-type="journal">R.S. Sutton, &quot;Reinforcement Learning: Past, Present and Future&quot;, Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), Vol. 1585, 1998, 195–197.</mixed-citation>
                    </ref>
                                    <ref id="ref2">
                        <label>2</label>
                        <mixed-citation publication-type="journal">L.P. Kaelbling, M.L. Littman, A.W. Moore, &quot;Reinforcement Learning:  A Survey&quot;,  J. Artif. Intell. Res., Vol. 4, 1996, pp. 237–285.</mixed-citation>
                    </ref>
                                    <ref id="ref3">
                        <label>3</label>
                        <mixed-citation publication-type="journal">R.S. Sutton, A.G. Barto, &quot;Reinforcement Learning: An Introduction&quot;, 1998.</mixed-citation>
                    </ref>
                                    <ref id="ref4">
                        <label>4</label>
                        <mixed-citation publication-type="journal">J. Xue, Q. Gao, W. Ju, &quot;Reinforcement learning for engine idle speed control&quot;, 2010 Int. Conf. Meas. Technol. Mechatronics Autom. ICMTMA 2010, Vol. 2, 2010, pp. 1008–1011.</mixed-citation>
                    </ref>
                                    <ref id="ref5">
                        <label>5</label>
                        <mixed-citation publication-type="journal">E. Uchibe, M. Asada, K. Hosoda, &quot;Behavior coordination for a mobile robot using modular reinforcement learning&quot;, IEEE Int. Conf. Intell. Robot. Syst., Vol. 3, 1996, pp. 1329–1336.</mixed-citation>
                    </ref>
                                    <ref id="ref6">
                        <label>6</label>
                        <mixed-citation publication-type="journal">Z. Linan, Y. Peng, C. Lingling, Z. Xueping, T. Yantao, &quot;Obstacle avoidance of multi mobile robots based on behavior decomposition reinforcement learning&quot;, 2007 IEEE Int. Conf. Robot. Biomimetics, ROBIO, 2007, pp. 1018–1023.</mixed-citation>
                    </ref>
                                    <ref id="ref7">
                        <label>7</label>
                        <mixed-citation publication-type="journal">N.J. Van Eck, M. Van Wezel, &quot;Application of reinforcement learning to the game of Othello&quot;, Comput. Oper. Res., Vol. 35, 2008, pp. 1999–2017.</mixed-citation>
                    </ref>
                                    <ref id="ref8">
                        <label>8</label>
                        <mixed-citation publication-type="journal">C.J.C.H. Watkins, &quot;Learning from delayed rewards&quot;, 1989.</mixed-citation>
                    </ref>
                                    <ref id="ref9">
                        <label>9</label>
                        <mixed-citation publication-type="journal">C.J.C.H. Watkins, P. Dayan, &quot;Q-learning&quot;, Mach. Learn. 1992, Vol. 83, 8, 1992, pp. 279–292,</mixed-citation>
                    </ref>
                                    <ref id="ref10">
                        <label>10</label>
                        <mixed-citation publication-type="journal">V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, M. Riedmiller, &quot; Playing Atari with Deep Reinforcement Learning&quot;, 2013.</mixed-citation>
                    </ref>
                                    <ref id="ref11">
                        <label>11</label>
                        <mixed-citation publication-type="journal">D. Silver, G. Lever, N. Heess, T. Degris, D. Wierstra, M. Riedmiller, &quot;Deterministic Policy Gradient Algorithms&quot;.</mixed-citation>
                    </ref>
                                    <ref id="ref12">
                        <label>12</label>
                        <mixed-citation publication-type="journal">T.P. Lillicrap, J.J. Hunt, A. Pritzel, N. Heess, T. Erez, Y. Tassa, D. Silver, D. Wierstra, &quot; Continuous control with deep reinforcement learning&quot;,  4th Int. Conf. Learn. Represent. ICLR 2016 - Conf. Track Proc., 2015.</mixed-citation>
                    </ref>
                                    <ref id="ref13">
                        <label>13</label>
                        <mixed-citation publication-type="journal">S. Fujimoto, H. Hoof, D. Meger, &quot;Addressing Function Approximation Error in Actor-Critic Methods&quot;, http://proceedings.mlr.press/v80/fujimoto18a.html, 2018.</mixed-citation>
                    </ref>
                                    <ref id="ref14">
                        <label>14</label>
                        <mixed-citation publication-type="journal">F. Harashima, S. Kondo, &quot;Design Method For Digital Speed Control System Of Motor Drives&quot;, PESC Rec. - IEEE Annu. Power Electron. Spec. Conf., 1982, pp. 289–297.</mixed-citation>
                    </ref>
                                    <ref id="ref15">
                        <label>15</label>
                        <mixed-citation publication-type="journal">D. Germanton, M. Lehr, &quot;Variable speed DC motor controller apparatus particularly adapted for control of portable-power tools&quot;, 1989.</mixed-citation>
                    </ref>
                                    <ref id="ref16">
                        <label>16</label>
                        <mixed-citation publication-type="journal">Y. Hoshino, &quot;A proposal of Reinforcement Learning System to Use Knowledge effectively&quot;, 2003, pp. 1582–1585.</mixed-citation>
                    </ref>
                                    <ref id="ref17">
                        <label>17</label>
                        <mixed-citation publication-type="journal">S.J. Russell, P. Norvig, &quot;Artificial Intelligence A Modern Approach&quot;, 2003.</mixed-citation>
                    </ref>
                                    <ref id="ref18">
                        <label>18</label>
                        <mixed-citation publication-type="journal">R.S. Sutton, D. Mcallester, S. Singh, Y. Mansour, &quot;Policy gradient methods for reinforcement learning with function approximation&quot;, Adv. NEURAL Inf. Process. Syst. 12, Vol. 12, 2000, pp. 1057--1063.</mixed-citation>
                    </ref>
                                    <ref id="ref19">
                        <label>19</label>
                        <mixed-citation publication-type="journal">H. van Hasselt, A. Guez, D. Silver, &quot;Deep Reinforcement Learning with Double Q-Learning&quot;, Proc. AAAI Conf. Artif. Intell. 30, 2016.</mixed-citation>
                    </ref>
                                    <ref id="ref20">
                        <label>20</label>
                        <mixed-citation publication-type="journal">W.B. Knox, P. Stone, &quot;Reinforcement learning from human reward: Discounting in episodic tasks&quot;, Proc. - IEEE Int. Work. Robot Hum. Interact. Commun., 2012, pp. 878–885.</mixed-citation>
                    </ref>
                                    <ref id="ref21">
                        <label>21</label>
                        <mixed-citation publication-type="journal">University of Michigan: Control Tutorials for MATLAB and Simulink - Motor Speed: System Modeling, https://ctms.engin.umich.edu/CTMS/index.php?example=MotorSpeed&amp;section=SystemModeling.</mixed-citation>
                    </ref>
                            </ref-list>
                    </back>
    </article>
