<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>LLM | Sen He</title><link>https://senhe.ai/tags/llm/</link><atom:link href="https://senhe.ai/tags/llm/index.xml" rel="self" type="application/rss+xml"/><description>LLM</description><generator>HugoBlox Kit (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Thu, 01 Jan 2026 00:00:00 +0000</lastBuildDate><image><url>https://senhe.ai/media/icon_hu_da05098ef60dc2e7.png</url><title>LLM</title><link>https://senhe.ai/tags/llm/</link></image><item><title>Harnessing large language models for virtual reality exploration testing: a case study</title><link>https://senhe.ai/publications/qi-2026-vr-llm/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://senhe.ai/publications/qi-2026-vr-llm/</guid><description/></item><item><title>Performance analysis of AI-generated code: A case study of Copilot, Copilot Chat, CodeLlaMa, and DeepSeek-Coder models</title><link>https://senhe.ai/publications/li-2026-ai-codegen/</link><pubDate>Thu, 01 Jan 2026 00:00:00 +0000</pubDate><guid>https://senhe.ai/publications/li-2026-ai-codegen/</guid><description/></item><item><title>AI-Driven Software Testing</title><link>https://senhe.ai/projects/ai-testing/</link><pubDate>Sat, 01 Jun 2024 00:00:00 +0000</pubDate><guid>https://senhe.ai/projects/ai-testing/</guid><description>&lt;h2 id="overview"&gt;Overview&lt;/h2&gt;
&lt;p&gt;Modern software systems — from virtual reality applications to microservice architectures — present unique testing challenges that traditional approaches struggle to address. This project explores how &lt;strong&gt;large language models and AI techniques&lt;/strong&gt; can be applied to automate and improve software testing across these domains.&lt;/p&gt;
&lt;h2 id="research-directions"&gt;Research Directions&lt;/h2&gt;
&lt;p&gt;&lt;strong&gt;LLM-Based VR Testing&lt;/strong&gt;: We investigate using LLMs for exploration testing of VR applications, where the state space is large and traditional test generation methods are insufficient. Our work demonstrates how LLMs can generate meaningful interaction sequences that achieve high code coverage.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Code Clone Detection in VR Software&lt;/strong&gt;: Empirical studies on code cloning patterns specific to VR software development, identifying unique maintenance and security implications.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Software Security in VR&lt;/strong&gt;: Investigating software security weaknesses across VR projects, examining when and why vulnerabilities emerge during the development lifecycle.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Microservice Log Analysis with AI&lt;/strong&gt;: Systematic analysis of how AI techniques can be applied to microservice log data for anomaly detection, root cause analysis, and system health monitoring.&lt;/p&gt;</description></item><item><title>Multi-Agent LLM Orchestration for Software Engineering</title><link>https://senhe.ai/projects/perforchestra/</link><pubDate>Mon, 01 Jan 2024 00:00:00 +0000</pubDate><guid>https://senhe.ai/projects/perforchestra/</guid><description>&lt;h2 id="overview"&gt;Overview&lt;/h2&gt;
&lt;p&gt;Large language models have demonstrated remarkable capabilities in code generation, but individual models often produce code with functional errors or performance inefficiencies. This project investigates &lt;strong&gt;multi-agent orchestration frameworks&lt;/strong&gt; that coordinate multiple LLMs through structured collaboration pipelines — combining specialized agents for categorization, code generation, debugging, and refinement.&lt;/p&gt;
&lt;h2 id="key-contributions"&gt;Key Contributions&lt;/h2&gt;
&lt;p&gt;Our flagship system, &lt;strong&gt;PerfOrch&lt;/strong&gt;, introduces a memory-augmented multi-agent architecture where:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;A &lt;strong&gt;Categorizing Agent&lt;/strong&gt; classifies programming tasks using a fixed category vocabulary to enable retrieval of relevant historical solutions.&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Generator&lt;/strong&gt; and &lt;strong&gt;Debugger Agents&lt;/strong&gt; collaborate through iterative cycles to produce functionally correct code.&lt;/li&gt;
&lt;li&gt;A &lt;strong&gt;Refinement Agent&lt;/strong&gt; optimizes code performance using aggregated insights from the orchestration&amp;rsquo;s memory module.&lt;/li&gt;
&lt;li&gt;The architecture leverages asymmetric aggregation strategies (product vs. sum) across different pipeline stages.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;We evaluate across 5 frontier LLMs and demonstrate significant improvements over single-model baselines on both correctness and runtime performance metrics.&lt;/p&gt;
&lt;h2 id="status"&gt;Status&lt;/h2&gt;
&lt;p&gt;Manuscript submitted to ACM Transactions on Software Engineering and Methodology (TOSEM).&lt;/p&gt;</description></item></channel></rss>