<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>https://www.scipedia.com/wd/index.php?action=history&amp;feed=atom&amp;title=Liang_et_al_2019a</id>
		<title>Liang et al 2019a - Revision history</title>
		<link rel="self" type="application/atom+xml" href="https://www.scipedia.com/wd/index.php?action=history&amp;feed=atom&amp;title=Liang_et_al_2019a"/>
		<link rel="alternate" type="text/html" href="https://www.scipedia.com/wd/index.php?title=Liang_et_al_2019a&amp;action=history"/>
		<updated>2026-04-21T17:48:14Z</updated>
		<subtitle>Revision history for this page on the wiki</subtitle>
		<generator>MediaWiki 1.27.0-wmf.10</generator>

	<entry>
		<id>https://www.scipedia.com/wd/index.php?title=Liang_et_al_2019a&amp;diff=194124&amp;oldid=prev</id>
		<title>Scipediacontent: Scipediacontent moved page Draft Content 514190107 to Liang et al 2019a</title>
		<link rel="alternate" type="text/html" href="https://www.scipedia.com/wd/index.php?title=Liang_et_al_2019a&amp;diff=194124&amp;oldid=prev"/>
				<updated>2021-01-28T20:27:03Z</updated>
		
		<summary type="html">&lt;p&gt;Scipediacontent moved page &lt;a href=&quot;/public/Draft_Content_514190107&quot; class=&quot;mw-redirect&quot; title=&quot;Draft Content 514190107&quot;&gt;Draft Content 514190107&lt;/a&gt; to &lt;a href=&quot;/public/Liang_et_al_2019a&quot; title=&quot;Liang et al 2019a&quot;&gt;Liang et al 2019a&lt;/a&gt;&lt;/p&gt;
&lt;table class=&quot;diff diff-contentalign-left&quot; data-mw=&quot;interface&quot;&gt;
				&lt;tr style='vertical-align: top;' lang='en'&gt;
				&lt;td colspan='1' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan='1' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;Revision as of 20:27, 28 January 2021&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan='2' style='text-align: center;' lang='en'&gt;&lt;div class=&quot;mw-diff-empty&quot;&gt;(No difference)&lt;/div&gt;
&lt;/td&gt;&lt;/tr&gt;&lt;/table&gt;</summary>
		<author><name>Scipediacontent</name></author>	</entry>

	<entry>
		<id>https://www.scipedia.com/wd/index.php?title=Liang_et_al_2019a&amp;diff=194123&amp;oldid=prev</id>
		<title>Scipediacontent: Created page with &quot; == Abstract ==   Traffic Management (ATM) is a complex decision-making process. Air traffic controllers’ decision on aircraft trajectory control actions directly leads to t...&quot;</title>
		<link rel="alternate" type="text/html" href="https://www.scipedia.com/wd/index.php?title=Liang_et_al_2019a&amp;diff=194123&amp;oldid=prev"/>
				<updated>2021-01-28T20:26:57Z</updated>
		
		<summary type="html">&lt;p&gt;Created page with &amp;quot; == Abstract ==   Traffic Management (ATM) is a complex decision-making process. Air traffic controllers’ decision on aircraft trajectory control actions directly leads to t...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;&lt;br /&gt;
== Abstract ==&lt;br /&gt;
&lt;br /&gt;
 Traffic Management (ATM) is a complex decision-making process. Air traffic controllers’ decision on aircraft trajectory control actions directly leads to the efficiency of traffic flow management. In the Automated Point Merge Trajectory Planning (APMTP) problem, it aims to realize an automated routine trajectory management in Terminal Manoeuvring Area (TMA) with an intelligent decision-making agent.An Artificial Intelligence-based approach, mainly Reinforcement Learning (RL) algorithm, is applied to adaptively and smartly integrate four types of de-conflict actions for solving conflicts with fewer delays on the environment. In this paper, we will mainly discuss the policy optimization in APMTP, focus on improving the agent’s learning quality and exploration efficiency. Firstly,application of RL in adaptive trajectory planning is presented.APMTP problem is adaptively divided into several sub-problems.For each sub-problem, an online policy π is applied to guide the simulation and optimization modules to find out the conflict free and less-delay solution. The online policy π is a scale of weight distribution for choosing desirable actions. It follows the rule of Roulette-wheel selection with weighted probability. The highest desirable decision variable has the largest share of the roulette wheel, while the lowest desirable decision variable has the smallest share of the roulette wheel. The RL direct policy optimization algorithm is designed to update the online policy π.Finally, experiments are built up for validation of the proposed policy optimization algorithm for the intelligent decision-making in APMTP. The results in the test environment show that learning agent with different exploration and exploitation ability will result in different system performance in conflict resolution and delay Refereed/Peer-reviewed&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Original document ==&lt;br /&gt;
&lt;br /&gt;
The different versions of the original document can be found in:&lt;br /&gt;
&lt;br /&gt;
* [http://dx.doi.org/10.13140/rg.2.2.30535.85926 http://dx.doi.org/10.13140/rg.2.2.30535.85926]&lt;br /&gt;
&lt;br /&gt;
* [https://hal-enac.archives-ouvertes.fr/hal-02267452/file/conference_041818.pdf https://hal-enac.archives-ouvertes.fr/hal-02267452/file/conference_041818.pdf]&lt;br /&gt;
&lt;br /&gt;
* [https://hdl.handle.net/11541.2/138969 https://hdl.handle.net/11541.2/138969]&lt;br /&gt;
&lt;br /&gt;
* [http://xplorestaging.ieee.org/ielx7/9069808/9081608/09081789.pdf?arnumber=9081789 http://xplorestaging.ieee.org/ielx7/9069808/9081608/09081789.pdf?arnumber=9081789],&lt;br /&gt;
: [http://dx.doi.org/10.1109/dasc43569.2019.9081789 http://dx.doi.org/10.1109/dasc43569.2019.9081789]&lt;br /&gt;
&lt;br /&gt;
* [https://hal-enac.archives-ouvertes.fr/hal-02267452 https://hal-enac.archives-ouvertes.fr/hal-02267452],&lt;br /&gt;
: [https://hal-enac.archives-ouvertes.fr/hal-02267452/document https://hal-enac.archives-ouvertes.fr/hal-02267452/document],&lt;br /&gt;
: [https://hal-enac.archives-ouvertes.fr/hal-02267452/file/conference_041818.pdf https://hal-enac.archives-ouvertes.fr/hal-02267452/file/conference_041818.pdf]&lt;br /&gt;
&lt;br /&gt;
* [https://hal-enac.archives-ouvertes.fr/hal-02267452/document https://hal-enac.archives-ouvertes.fr/hal-02267452/document],&lt;br /&gt;
: [https://hal-enac.archives-ouvertes.fr/hal-02267452 https://hal-enac.archives-ouvertes.fr/hal-02267452],&lt;br /&gt;
: [https://academic.microsoft.com/#/detail/2979872972 https://academic.microsoft.com/#/detail/2979872972]&lt;br /&gt;
&lt;br /&gt;
* [ ]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
DOIS: 10.1109/dasc43569.2019.9081789 10.13140/rg.2.2.30535.85926 10.1109/dasc43569.2019.9081789.&lt;/div&gt;</summary>
		<author><name>Scipediacontent</name></author>	</entry>

	</feed>