<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>3 | Academic</title><link>https://qiyuhan.netlify.app/publication-type/3/</link><atom:link href="https://qiyuhan.netlify.app/publication-type/3/index.xml" rel="self" type="application/rss+xml"/><description>3</description><generator>Wowchemy (https://wowchemy.com)</generator><language>en-us</language><lastBuildDate>Wed, 21 Dec 2022 00:00:00 +0000</lastBuildDate><image><url>https://qiyuhan.netlify.app/media/icon_hu0b7a4cb9992c9ac0e91bd28ffd38dd00_9727_512x512_fill_lanczos_center_3.png</url><title>3</title><link>https://qiyuhan.netlify.app/publication-type/3/</link></image><item><title>Online Statistical Inference for Matrix Contextual Bandit.</title><link>https://qiyuhan.netlify.app/publication/conference-paper-1/</link><pubDate>Wed, 21 Dec 2022 00:00:00 +0000</pubDate><guid>https://qiyuhan.netlify.app/publication/conference-paper-1/</guid><description>&lt;div class="alert alert-note">
&lt;div>
Click the &lt;em>link&lt;/em> below to view our work published in arxiv.
&lt;/div>
&lt;/div>
&lt;p>&lt;a href="https://arxiv.org/abs/2212.11385" target="_blank" rel="noopener">View our work&lt;/a>&lt;/p></description></item><item><title>Active Learning in RL with Human Feedback</title><link>https://qiyuhan.netlify.app/project/project-2/</link><pubDate>Tue, 21 Dec 2021 00:00:00 +0000</pubDate><guid>https://qiyuhan.netlify.app/project/project-2/</guid><description>&lt;p>In the realm of RL, the incorporation of human feedback serves as a framework for enhancing the agent&amp;rsquo;s learning process through domain-specific insights. While conventional RL algorithms strive to maximize reward by interacting with their environment, defining what precisely constitutes a &amp;ldquo;reward&amp;rdquo; is often a complex task. Such definitions demand relevant domain knowledge to appropriately specify &amp;ldquo;good&amp;rdquo; and &amp;ldquo;bad&amp;rdquo; agent behaviors. Moreover, human feedback often plays a critical role in establishing action constraints to ensure safety and fairness. Research in this domain shows significant promise for enhancing the adaptability of RL algorithms across a wide range of real-world applications. However, there is still a lack of comprehensive scientific investigations in this area. As such, one of my forthcoming research objectives is to contribute substantively to the development of this field.&lt;/p></description></item><item><title>Off-Policy Evaluation For Low-Rank Tensor Markov Decision Processes.</title><link>https://qiyuhan.netlify.app/publication/conference-paper-2/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://qiyuhan.netlify.app/publication/conference-paper-2/</guid><description/></item><item><title>Statistical Inference for Contextual Bandit with Delayed Action Effects</title><link>https://qiyuhan.netlify.app/project/project-1/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>https://qiyuhan.netlify.app/project/project-1/</guid><description>&lt;p>In many real-world situations, the consequences of actions are not immediately apparent and may only become observable after a certain time has elapsed. For instance, in healthcare settings where medications are administered to patients, the drug&amp;rsquo;s effects are not instantly evident. Often, the current reward or outcome is connected to past actions. Although some existing literature has addressed this temporal aspect by considering time-discounted effects or delayed consequences, the role of statistical inference becomes crucial when it comes to sensitive applications like medication administration. Extra caution is required before making any recommendations regarding drug intake, highlighting the need for rigorous statistical evaluation. The application-driven motivation for this line of research can be rooted in the study of medication effects for diabetes, where the historical medical treatment has prolonged effects on the current patients&amp;rsquo; well-being. A significant challenge in this context is the violation of the Markov assumption, due to the extended impact of treatment actions. Within the scope of this study, my objective is to develop a methodology that transforms conventional problem settings, predominantly centered around off-policy evaluation, into Markov Decision Processes. This transformation will facilitate the application of established RL algorithms to this domain.&lt;/p></description></item></channel></rss>