Oms On The Fly Multi Objective Self Reflective Ad Keyword Generation Via Llm Agent
Abstract
Section titled “Abstract”Keyword decision in Sponsored Search Advertising is critical to the success of ad campaigns. While LLM-based methods offer automated keyword generation, they face three major limitations: reliance on large-scale query-keyword pair data, lack of online multi-objective performance monitoring and optimization, and weak quality control in keyword selection. These issues hinder the agentic use of LLMs in fully automating keyword decisions by monitoring and reasoning over key performance indicators such as impressions, clicks, conversions, and CTA effectiveness. To overcome these challenges, we propose OMS, a keyword generation framework that is On-the-fly (requires no training data, monitors online performance, and adapts accordingly), Multi-objective (employs agentic reasoning to optimize keywords based on multiple performance metrics), and Self-reflective (agentically evaluates keyword quality). Experiments on benchmarks and real-world ad campaigns show that OMS outperforms existing methods; Ablation and human evaluations confirm the effectiveness of each component and the quality of generated keywords.
1 Introduction
Section titled “1 Introduction”In Sponsored Search Advertising (SSA) (Fain and Pedersen, 2006), advertisers participate in bidding when their keyword lists broadly match the user’s query. Figure 1 shows how ad keywords affect the SSA process. When a user enters a search query, the platform runs a real-time auction. Advertisers’ keyword lists decide whether their ads appear, where they are placed, and how likely users are to click or convert. This highlights that while impression opportunities may be broadly shared, conversion likelihood heavily depends on how well an advertiser’s keyword list (Wu et al., 2009; Chen
et al., 2008) captures the user’s intent, making keyword setting a critical factor in SSA.
Even today, automating keyword decisions beyond SSA platforms like Google, Bing (Google, 2025; Microsoft, 2025) remains challenging, as only these platforms have access to large-scale query–keyword data. Companies typically rely on keyword suggestion tools of SSA platforms or specialized ad agencies. However, even for SSA platforms and agencies, selecting effective keywords still requires ongoing effort—monitoring performance, updating keywords, and investing time and budget (Wang et al., 2025a) to keep up with changing market trends and improve campaign outcomes.
In academic research, early methods used generative models such as GANs (Lee et al., 2018) and LSTMs (Lian et al., 2019) to generate keywords under this research direction. Recently, the emergence of large language models (LLMs) has helped overcome data scarcity, enabling advertisers to generate their own keywords. In this line of research, LKG (Wang et al., 2024) fine-tuned an LLM with constrained decoding to generate SSA keywords from real user queries. Meanwhile, OKG (Wang et al., 2025a) introduces a keyword generation agent that leverages the reasoning ability of a ReAct-style LLM agent (Yao et al., 2023) to incorporate click feedback into the generation process, whose agentic reasoning helps identify underperforming keywords, missed intents, and emerging trends.
Despite these advancements, existing SSA keyword generation methods still face several limitations, as summarized in Table 1. Traditional approaches such as WIKG (Nie et al., 2019), GAN-based (Lee et al., 2018), and LSTM-based (Lian et al., 2019) methods are tightly dependent on large-scale data to train the models, and lack real-time performance monitoring and multi-objective optimization abilities. LKG (Wang et al., 2024), while leveraging LLMs trained on a static keyword
* indicates equal contribution. Bowen Chen completed this work during his internship at Sony. Corresponding author: Zhao.Wang@sony.com

Figure 1: An illustrative example: among the three advertisers (blue blocks and yellow block) that received impressions, only one (yellow block) achieved a conversion.
dataset, lacks support for real-time feedback or reasoning during keyword generation, limiting its ability to adapt to rapid changes in user behavior. OKG (Wang et al., 2025a) has two major limitations. First, its decision-making is coarsegrained—it relies on predefined categories and aggregated click statistics, without jointly considering other important metrics such as conversions, CTA rates, and more. Second, its agentic process is static, simply following prompt-then-generate flow, which often results in low-quality keyword outputs.
We propose OMS, an On-the-fly, Multi-Objective and Self-Reflective ad keyword generation framework to address the above limitations and realize the functionalities outlined in Table 1. Specifically, our contributions are three-fold:
- We propose an agentic clustering-ranking module that monitors and clusters the currently active keywords, computes multi-objective scores, and ranks keywords both intra- and interclusters to identify high-impact keywords.
 - We introduce a multi-turn generationreflection module. New keywords are generated through agentic reasoning and iteratively refined based on self-reflective feedback.
 - Extensive Experimental Results demonstrate that OMS outperforms multiple baselines on both established benchmarks and real-world online ad campaigns. Ablation studies highlight the effectiveness of each component, and human preference evaluations confirm OMS’s superiority across multiple metrics.
 
2 Related Works
Section titled “2 Related Works”2.1 LLM Agents in Business Applications
Section titled “2.1 LLM Agents in Business Applications”LLM agents have demonstrated their potential in business areas, including financial question answer-
| Method | Capabilities | ||||||
|---|---|---|---|---|---|---|---|
| Free to Train  | Real-Time Monitoring  | Multi-Objective Optimization  | Self-Reflective Quality Control  | ||||
| WIKG | Х | Х | × | × | |||
| GAN-based | X | X | X | X | |||
| LSTM-based | X | X | X | X | |||
| LKG | X | X | ✓ | X | |||
| OKG | / | / | X | X | |||
| Ours | 1 | ✓ | ✓ | ✓ | 
Table 1: Capability comparison between the proposed OMS method and conventional methods for the SSA keyword generation.
ing (Fatemi and Hu, 2024), financial sentiment analysis (Xing, 2025), customer support (Zhong et al., 2023), outperforming conventional methods.
While LLM agents have shown strong performance across many business domains, most existing LLM-based approaches in advertising have not yet embraced the agentic paradigm. Recent studies have explored the use of LLMs for ad text generation (Mita et al., 2024; Wang et al., 2025b) and ad image creation (Chen et al., 2025), but these methods typically either do not leverage the full reasoning capabilities of LLM agents (Chen et al., 2025), or lack adaptive generation for real-time business performance (Mita et al., 2024; Noy and Zhang, 2023). As a result, such approaches remain confined to static, offline settings and struggle to cope with the dynamic, fast-changing nature of realworld advertising campaigns (Wang et al., 2025a), where factors such as brand reputation, product quality, and market trends evolve constantly.
2.2 Keyword Generation in SSA
Section titled “2.2 Keyword Generation in SSA”Early research utilizes knowledge graphs like Wikipedia (Nie et al., 2019) to generate seed keywords and expand those keywords by statistical information or concept hierarchy-based expansions (Joshi and Motwani, 2006; Wu et al., 2009; Chen et al., 2008). Later studies used GAN and LSTM models to generate keywords from user queries (Lee et al., 2018; Lian et al., 2019), but these relied heavily on proprietary data, restricting advertisers’ ability to customize keyword strategies. Recently, LLMs with their strong text generation ability have provided advertisers with opportunities to generate keywords. Wang et al. (2024) fine-tuned LLMs using query-keyword with click data and used beam-search to generate keywords. OKG (Wang et al., 2025a) further advanced this field by integrating product information and real-time
click performance data, enabling partial real-time adaptation. However, as detailed in Section 1, the click aggregation based optimization and promptthen-generate static agentic flow often results in low-quality keyword outputs.
3 Problem Statement
Section titled “3 Problem Statement”Given a product R advertised on a SSA platform with product information r, at time step t during the ad compagin period, we have the historical keywords up to time t, where I is the total number of historical keywords. Each keyword (used at any time step ) has a performance vector , consisting of some relevant metrics such as Click, Conversion and etc.
Given and the associated performance vectors , a new keyword set is generated at each time step t by a generation algorithm g:
(1)
The objective over the entire advertising campaign period T is to maximize the cumulative performance of all selected keywords:
s.t.
is the weight vector over performance metrics; B is the total budget available for the entire campaign. N is the maximum number of keywords that can be generated at each time step t.
4 Methodology
Section titled “4 Methodology”Figure 2 shows the overall OMS workflow process. The first module, Agentic Clustering-Ranking (Steps 2–4 in Figure 2), analyzes keyword intent, calculates multi-objective TOPSIS scores, and ranks keywords within and across clusters (Section 4.1). The second module, Multi-Turn Generation-Reflection (Steps 5–7), dynamically formulates prompts, generates new keywords using external tools, iteratively refines them based
on quality feedback and alysis, and re-clusters new generated keywords (Section 4.2). 2
4.1 Agentic Clustering-Ranking Module
Section titled “4.1 Agentic Clustering-Ranking Module”4.1.1 Keyword Intent Analysis
Section titled “4.1.1 Keyword Intent Analysis”At each time step t, we are given the current set of active (being set up in SSA) keywords , along with their corresponding cluster assignments . is calculated from the previous step t-1 by using eq.8. Each cluster contains a semantically coherent subset of keywords. Given the product description r and keywords in a cluster , OMS agent outputs an intent summary this cluster:
Here, LLM_Intent is a prompt template used to acquire keyword intent. The resulting provides an explanation of the likely search or marketing intent behind each cluster, such as emphasizing product features (e.g., “high resolution”) or addressing consumer concerns (e.g., “cost performance”). This intent summary enables our agent to reason about the current keyword setup and identify potential areas that may benefit from keyword expansion.
4.1.2 TOPSIS Score Calculation and Ranking
Section titled “4.1.2 TOPSIS Score Calculation and Ranking”To prioritize keywords based on multiple performance metrics, we compute a TOPSIS score for each active keyword (Zavadskas et al., 2006). Each performance vector is first normalized using min-max scaling, where positive metrics (e.g., Clicks, Conversions) and negative metrics (e.g., Cost) are handled accordingly. Let denote the normalized value of keyword for performance metric . The weight vector reflects the importance of each metric and can be predefined as hyperparameter or automatically adapted.
For each keyword with its normalized performance vector , we calculate its TOPSIS score (Hwang and Yoon, 1981) based on its relative distance to the ideal (all-ones) and anti-ideal (all-zeros) coordinates:
(4)
<sup>1Advertisers set the budget. SSA restricts the number of deployable keywords. The budget usage, like bidding strategy for keywords, is managed automatically by the SSA platform.
<sup>2All the prompts we used in this section can be found in Appendix D due to space limitations.

Figure 2: Overview of the OMS workflow. The process consists of two main modules: Agentic Clustering-Ranking (Steps 2–4) and Multi-Turn Generation-Reflection (Steps 5–7). The system monitors keyword performance (Step 1), ranks keywords based on multi-objective analysis (Steps 2–4), generates new candidates, iteratively refines them and re-clusters the new generated keywords (Steps 5–7) before deployment (Steps 8).
Entropy-based Weighting. If weights are not predefined, we compute them using entropy. Let be the normalized contribution of keyword to metric d. The entropy-based weight for each metric is then computed as:
(5)
Metrics with higher variance (lower entropy) receive larger weights, as they contain more potential space for keyword performance improvement.
Intra- and Inter-Cluster Ranking. For each cluster , we perform intra-cluster ranking by sorting the keywords in descending order of their TOPSIS scores . This reveals the most and least effective keywords within each intent group and supports keyword-level refinement.
We also compute a cluster-level score to enable inter-cluster ranking across different keyword groups: . Clusters are
then ranked by their scores to prioritize highperforming groups for expansion and flag underperforming ones for improvement.
4.2 Multi-Turn Generation-Reflection Module
Section titled “4.2 Multi-Turn Generation-Reflection Module”Dynamic Cluster Expansion for Many-Shot Prompting. To adaptively construct keyword generation prompts, we adaptively formulate a ranking-aware prompt LLM_Rank based on the cluster-level TOPSIS score and a predefined threshold . The prompt for each cluster is constructed as the following:
where is the intent summary of cluster from eq.3, and is the TOPSIS score of keyword . The function extracts both the highest-and lowest-ranked keywords for a bad performance
| Tool | Description | 
|---|---|
| Search Tool | Uses SerpAPI3 to retrieve product information via Google queries until enough data is deemed by the Agent to be gathered.  | 
| Reject Reflection | Analyzes poor-performing keywords and provides feedback to the Keyword Generator. | 
| Keyword Filter Tool | Flags validated low-performance or deployed keywords and asks the generator to replace them. | 
| Search Volume Validator | Calling API to check if keywords meet search volume thresholds to avoid automatic rejection by the SSA platform. | 
| Keyword Lexical Analysis | Detects common lexical patterns in SSA rejected keywords and provides an analysis report to avoid similar keywords. | 
| Category Analysis | Identifies clusters with high rejection rates and prompts the generator to replace keywords of the entire cluster. | 
Table 2: Tools integrated into the OMS keyword generation framework.
cluster while all keywords information are provided in a good performance cluster.
This prompt expands high-performing clusters to contribute rich, many-shot context, while summarizing low-performance clusters with contrastive examples to guide generation direction.
Tool-Return-Value-Guided Generation Flow.
Section titled “Tool-Return-Value-Guided Generation Flow.”Using the adaptive prompt above, OMS adopts a tool-return value-guided generation flow. Rather than letting the LLM autonomously decide which tool to invoke. Tools like the search volume checker, lexical analyzer, or filter (see Table 2) return explicit signals that guide the next step.
For instance, if some keywords are flagged for low historical conversion, this tool will ask to regenerate them. If the SSA API flags low search volume keywords, it informs the OMS to call the lexical tool to analyze their lexical features. This multi-turn process (Figure 2, Step 5) improves interpretability and control. 4
Keyword Quality Reflection. After generation, OMS performs a reflection step to evaluate the quality of each keyword ki ∈ Kt , using the product information r and peer set Kt . This process is formulated as following:
(7)
The LLM_Reflect returns feedback (for example, “redundant”, “irrelevant to product”), and suggestions regarding whether to keep or regenerate certain keywords. This cycle continues until the agent internally deems the keyword set Kt complete.
LLM-Assisted Re-Clustering. Finally, OMS reclusters the full keyword set K≤t ∪ Kt using Affinity Propagation (Frey and Dueck, 2007), resulting in updated clusters Ct = {C1, C2, … , CJ }. For each new keyword ki ∈ Kt , we identify its top-3 closest clusters in embedding space, denoted C top ki . A final assignment is determined by:
(8)
LLM_Assign selects the best-fit cluster or creates a new one to ensure semantic coherence for new keywords. This step is meaningful because keywords are often short and mutually similar, so relying solely on embedding distance can reduce cluster quality, degrading the performance at the next step by introducing semantically unrelated keywords for the intent analysis for clusters.
5 Experiments
Section titled “5 Experiments”LLM Backbone and Baselines. We use GPT-4o (OpenAI, 2024) as the backbone of OMS with temperature set to 0. Only in LLM_Reflect, we use o3 as the reflection LLM to leverage its stronger reasoning ability. In our experiments, we compare the OMS with the following baselines:
Google KW (Google, 2024): The built-in keyword planner tool provided by Google Ads.
GPT Series (OpenAI, 2024): GPT-4o, GPT-4.1, and o3 models prompted with the same product information and instruction to generate keywords. OKG (Wang et al., 2025a): We include the original OKG and Many-Shot OKG, which augments OKG with examples, e.g, prompting each keyword with its performances on each metric.
Metrics. We use following metrics:
ROUGE-1 (Lin, 2004) compares the words overlap between generated keywords with the product information to compare the lexical coverage.
BERTScore (Zhang et al., 2020) considers both semantic and lexical similarity between generated keywords and the provided product information.
Click, Search Volume, Cost per Click (CpC), and Competitor Score are four keyword performance metrics: Click and Search Volume represent the number of campaign clicks and user searches for keywords, CpC indicates keyword efficiency, and Competitor Score (judged by the SSA platform) reflects keywords’ competitiveness.
4Details of each tool are in the Appendix Sec A.2.

Figure 3: Performance Results on Benchmark Dataset. CpC is presented reversely, meaning a higher value means higher performance. Comp Score means the competitor score. The Rouge-1 and BERTScore are presented with raw values as their original values are between 0-1. Higher values indicate higher performance for all metrics.
Offline Benchmark Dataset. We use the benchmark released by OKG (Wang et al., 2025a). It collects keyword performance in the real world for 10 products with 1,000 keywords for each product.
Online A/B Test. We deploy OMS and OKG in Google Ads5 to conduct a real-world ad campaign.
Experiment Settings. For the offline benchmark test, keywords are generated over a time horizon of T = 5. At each time step t, keywords are adaptively generated by allowing OMS to observe performance feedback. For online A/B test, keywords are generated over a time horizon of T = 30Days, and we generate keywords every 3 days and collect the performance for generation and evaluation.
5.1 Offline Benchmark Results
Section titled “5.1 Offline Benchmark Results”We conducted experiments to compare the OMS with other baselines in the benchmark data. The visualization results of all products are in Figure 3 with normalized numerical values in Table 1. From this table, we can obtain:
(I) The OMS method outperforms baseline methods in all metrics, showing the effectiveness of the proposed method. The Many-Shot version OKG shows complicated results. It outperforms OKG in some metrics while showing lower performance in other metrics, showing its instability in learning from examples.
| Method | Clicks↑ CpC↓ | Search Volume↑  | Comp | Score↑ Rouge-1↑ | BERT Score ↑  | |
|---|---|---|---|---|---|---|
| Google KW | 0.70 | 1.00 | 0.62 | 0.73 | 0.16 | 0.41 | 
| GPT-4o | 0.74 | 0.95 | 0.79 | 0.77 | 0.18 | 0.54 | 
| o3 | 0.74 | 0.95 | 0.59 | 0.77 | 0.18 | 0.54 | 
| GPT-4.1 | 0.80 | 0.90 | 0.80 | 0.70 | 0.18 | 0.54 | 
| OKG | 0.85 | 0.86 | 0.88 | 0.83 | 0.21 | 0.55 | 
| + Many-Shot | 0.89 | 0.88 | 0.82 | 0.72 | 0.20 | 0.56 | 
| OMS | 0.92 | 0.84 | 0.90 | 0.89 | 0.23 | 0.62 | 
Table 3: Averaged normalized results of the benchmark.
- (II) The naive Google KW, GPT-4o, GPT-4.1, and o3 show a lower performance compared to both OKG and OMS, indicating that general solutions are not yet suitable for a highly customized situation like SSA keyword generation, showing the necessity of developing an agentic framework.
 - (III) The OMS has a higher performance in the Rouge-1 and BERTScore. This shows that the generated keywords are more related to the product information than keywords generated by other baselines, meaning the generated keywords may attract customers with higher purchase potential.
 
5.2 Effects of Shots and Rounds
Section titled “5.2 Effects of Shots and Rounds”We analyzed how the number of few-shot examples and generation rounds (T) affect performance, demonstrating OMS’s ability to utilize examples and maintain keyword quality over time.
Number of Shots (I) For Many-shot OKG, performance is unstable and drops when using over
5 https://ads.google.com/home/

Figure 4: Performance change with Number of Shots and Generation Rounds (Under 80 examples).
| Method | Clicks↑ CpC↓ | Search Volume↑  | Comp | Score↑ Rouge-1↑ | BERT Score ↑  | |
|---|---|---|---|---|---|---|
| OMS | 592 | 0.16 | 2866 | 92.7 | 0.25 | 0.63 | 
| Agentic Clustering-Ranking | ||||||
| -w/o Intent | 420 | 0.17 | 1166 | 88.7 | 0.17 | 0.58 | 
| -w/o TOPSIS | 328 | 0.28 | 866 | 84.3 | 0.14 | 0.56 | 
| Generation and Multi-turn Reflection | ||||||
| -w/o Reflection | 431 | 0.18 | 1923 | 90.3 | 0.16 | 0.56 | 
| -w/o LAR | 341 | 0.19 | 1324 | 90.1 | 0.22 | 0.60 | 
| Single Objective Optimization | ||||||
| OMS Click | 778 | 0.21 | 1423 | 91.7 | 0.20 | 0.59 | 
| OKG Click | 638 | 0.22 | 1823 | 90.7 | 0.20 | 0.62 | 
| OMS CpC | 451 | 0.14 | 1889 | 89.6 | 0.21 | 0.60 | 
| OKG CpC | 348 | 0.16 | 1903 | 90.1 | 0.22 | 0.58 | 
Table 4: Ablation Study over Alpha Camera product. Intent means Intention Analysis. LAR means the LLM Assisted Re-Clustering.
100 examples, likely due to noise in individual keywords that the LLM cannot effectively learn from. (II) In contrast, OMS improves as more examples are added. While its performance slightly drops when the example pool is small, as meaningful clusters are not been constructed yet. After the accumulation of example pools, we see the performance increases as clustered keyword examples with intention analysis could filter noise and provide more informative generation direction.
Number of Rounds (I) Both methods are augmented with 80 examples, and both methods improve over multiple rounds, but OMS improves faster and reaches a higher performance ceiling. (II) Multishot OKG suffers from semantic drift as it does not evaluate keyword quality, reflected by declining BERTScore and Rouge-1 with the increase of rounds. However, OMS maintains a
strong relation to the product across rounds due to the self-reflection over keyword quality.
5.3 Abalation Study
Section titled “5.3 Abalation Study”We conducted an ablation study to assess the importance of each OMS component with results in Table 4. We can obtain that:
- (I) Agentic Clustering-Ranking: Removing either intent analysis or the TOPSIS score reduces performance. Particularly, removing TOPSIS by prompting with raw, unprocessed metric values significantly degrades results. Without a unified performance objective, the LLM struggles to make decisions, especially under multiple metrics, and even fails to optimize a single metric among them.
 - (II) Multi-Turn Reflection: Disabling the reflection step results in lower Rouge-1 and BERTScore, indicating weaker alignment with product content. Removing LOC-based clustering and relying only on embeddings also hurts performance, as short, similar keywords form low-quality clusters that mislead the generation process.
 - (III) Single-Objective Optimization: Optimizing only one metric can improve that metric, but overall performance drops. OMS still outperforms OKG even under single-objective settings.
 
5.4 Human Preference and Case Study
Section titled “5.4 Human Preference and Case Study”To evaluate human preference, we anonymized the keywords generated by all methods and asked three professional annotators to rate them (1–5) on coverage, relevance, specialty, redundancy, alignment with user search behavior, and overall quality across five products. Results are shown in Table 6,
Sony α-series mirrorless cameras use full- frame & APS-C bodies with breakthrough stacked global shutter sensors. The flagship α9 III captures action at 1/80 000 s and up to 120 fps blackout-free bursts, freezing decisive moments without distortion. A powerful BIONZ XR engine and AI processing unit enable real-time tracking AF that recognizes people, animals, birds, insects, and vehicles with skeletal awareness. The diverse lineup spans the high-resolution 61 MP α7CR, wildlife-focused α7R V, and the low-light α7S III, plus the APS-C α6700 with enhanced eye-AF. Creators can choose from 70+ E-mount lenses—including the lightweight FE 24-50 mm F 2.8 G—for versatile, high-quality imaging. Workflow is streamlined via Creators’ Cloud integration, allowing direct uploads to Google Drive and Adobe Photoshop, while select bodies feature C2PA authenticity signatures to guarantee tamper-proof journalism. Together, the α-series unleashes high-speed capture and seamless AI-era production for professionals and creators alike.
GPT-4: Sony α camera, α9 III mirrorless, α7 IV full-frame, AI tracking, High-speed shutter camera, BIONZ XR image processing, Sports photography camera, High quality Sony Camera
GPT-4.1:Animal photography AI focus, Bird and landscape photography, Video production camera, AI-powered camera, AI camera with automatic recognition, Photo cloud integration, Smooth editing camera, G Master lens
o3: Sony α9 III, Blackout-free shooting, α 7 CR 61 MP, α 7 R V 61 MP, High-resolution mirrorless, α 7 S III for video, α 7 S III high sensitivity, Low-light shooting camera, Sony FE 24-50 mm F2.8, Lightweight E-mount lens
OKG: Sony flagship α camera, High-speed shooting camera, High-resolution camera, High-res landscape camera, α 9 III high-speed shutter, Instant capture camera, Real-time tracking camera, Compact lightweight camera, Noise-reduction capable camera, Automatic photo upload camera, Authenticity-signature camera
OMS:High-speed shutter camera, 120 fps α 9 III, Mirrorless sports camera, Real-time tracking AF, Eye AF camera, 61 MP α 7 CR, Low-noise video recording, Creators’ Cloud integration, Direct upload to Photoshop, Authenticity-signature support
Table 5: Case study for the Sony Neural Network Console Product with core features labeled with different colors.
| Method Cove↑ Rele↑ Spec↑ Redund↓ Align↑ Overall↑ | ||||||
|---|---|---|---|---|---|---|
| GPT-4o | 3.2 | 3.0 | 3.2 | 3.8 | 3.0 | 2.6 | 
| GPT-4.1 | 3.2 | 3.1 | 3.2 | 3.8 | 3.0 | 2.6 | 
| o3 | 3.1 | 2.9 | 3.4 | 3.3 | 2.8 | 3.1 | 
| OKG | 3.2 | 3.4 | 3.4 | 3.3 | 3.3 | 3.5 | 
| OMS | 3.5 | 3.8 | 3.9 | 3.0 | 3.9 | 3.7 | 
Table 6: Human Preference Ranking for Generated Keywords. Cove, Rele, Spec, Redund, Align means Coverage, Relevance, Speciality, Redundancy, Alignment.
with a case study in Table 5.
- (I) Human Preference: OMS receives the highest scores across all metrics. GPT-4o and GPT-4.1 often generate redundant or generic keywords. The o3 model generates more specialized keywords but lacks alignment with real search behavior. OKG performs better than GPT models but was still outperformed by OMS, especially in relevance, alignment, and specialty. OMS also slightly leads in coverage, capturing more core product features.
 - (II) Case Study: OMS covers the most product features among all methods. In contrast, o3 overfocuses on niche series names, GPT-4o/4.1 produces overly general (High quality Sony camera) or off-target keywords (Smooth editing), and OKG had similar issues. OMS generates keywords that were both relevant and specific, balancing product detail and user intent effectively.
 
5.5 Online A/B Test on Google Ads
Section titled “5.5 Online A/B Test on Google Ads”We deployed OMS and OKG in a real Google Ads campaign using an A/B test with the same budget for a campaign period, with results in Table 7. 6
(I) Performance Metrics: OMS outperforms OKG across all performance metrics, especially in
| Metric | OMS | OKG | Relative Gain | 
|---|---|---|---|
| Performance Related Metrics (Higher is Better)↑ | |||
| Conversion | 33 | 29 | +13.8% | 
| Clicks | 724 | 713 | +1.5% | 
| Impression | 11,586 | 11,422 | +1.4% | 
| CTR | 6.25% | 6.24% | +0.2% | 
| C.V Rate | 4.56% | 4.07% | +12.0% | 
| Cost Related Metrics (Lower is Better)↓ | |||
| CPA | ¥2,805 | ¥3,192 | -12.1% | 
| CPC | ¥128 | ¥130 | -1.5% | 
Table 7: A/B Test Performance Comparison (OMS vs. OKG). CTR means Click Through Rate, indicating the chance of clicking the campaign advertisement. C.V Rate means Click-to-Conversion rate, indicating the chance of conversion after clicking. CPA means the cost per conversion. The cost unit is Japanese Yen.
Cost ¥92,569 ¥92,575 -0.01%
Conversion and C.V Rate, which are key indicators in real-world advertising, indicating that keywords from OMS are more attractive and focused.
(II) Cost Efficiency: Under the same budget, OMS achieves lower CPA and CpC, meaning it obtained more clicks and conversions at a lower cost. In SSA bidding, even a slightly higher bid from other advertisers can dominate other keywords in SSA results. However, OMS still achieves better results with lower cost, highlighting its real-world efficiency and effectiveness.
6 Conclusion
Section titled “6 Conclusion”In this study, we proposed OMS—a SSA keyword generation framework that is On-the-fly (requires no offline training, adapts to real-time performance), Multi-objective (optimizes multiple metrics with dynamic prompting), and Self-reflective
6Details of the A/B test are in the Appendix Sec A.1.1.
(agentically evaluates keyword quality). Experiments show that OMS outperforms prior methods across diverse product benchmarks, with stable performance scaling in both shot count and generation rounds. Ablation study analyzed the importance of each component. Human preference and online A/B test confirm its real-world effectiveness.
7 Limitation
Section titled “7 Limitation”While we would like to implement methods that are trained on the query-keyword pair to generate keywords, and compare the proposed method with them, most of those datasets are not released, as the query information in the SSA platform is the in-house data of each platform, which links with personal privacy. Previous methods did not release such data, and thus, we are not able to train such a model. However, such methods are provided as a tool or API, like the Google Keyword Planner Service or Bing Keyword Planner Service. Though the technical details of such a keyword planner service are not released and provided in its documentation, we treat this method as the representative of traditional methods for SSA keyword generation that require query-keyword pair generation.
While we conducted an online A/B test to compare the OKG and OMS, we were not able to extend it into an A/B/n testing to compare multiple methods in a real-world campaign due to budget limitations. However, as OMS consistently outperforms other baselines both in the benchmark, which covers products of different popularity, and in the online A/B tests, we think OMS could outperform other baseline methods in most situations.
8 Ethical Considerations
Section titled “8 Ethical Considerations”For the benchmark we used, we have ensured the usage aligns with the data license and its intended usage. The collected keywords performance during an A/B test does not contain any personal information, as it is intended for a product campaign set by the company, in which we are only able to see the performance and are not able to see who contributed to a certain conversion or click. Therefore, this research work is not concerned with ethical issues.
For the usage of AI tools, we used ChatGPT for polishing the writing of this paper. The usage of AI tools is not beyond this scope.
References
Section titled “References”- 
Xingye Chen, Wei Feng, Zhenbang Du, Weizhen Wang, Yanyin Chen, Haohan Wang, Linkai Liu, Yaoyu Li, Jinyuan Zhao, Yu Li, et al. 2025. Ctr-driven advertising image generation with multimodal large language models. In Proceedings of the ACM on Web Conference 2025, pages 2262–2275.
 - 
Yifan Chen, Gui-Rong Xue, and Yong Yu. 2008. Advertising keyword suggestion based on concept hierarchy. In Proceedings of the 2008 International Conference on Web Search and Data Mining, WSDM ‘08, page 251–260, New York, NY, USA. Association for Computing Machinery.
 - 
Daniel C. Fain and Jan Pedersen. 2006. Sponsored search: A brief history. Bulletin of The American Society for Information Science and Technology, 32:12– 13.
 - 
Sorouralsadat Fatemi and Yuheng Hu. 2024. Enhancing financial question answering with a multi-agent reflection framework. In Proceedings of the 5th ACM International Conference on AI in Finance, ICAIF ‘24, page 530–537. ACM.
 - 
Brendan J. Frey and Delbert Dueck. 2007. Clustering by passing messages between data points. Science, 315(5814):972–976.
 - 
Google. 2024. Google keyword planner. https://ads. google.com/home/tools/keyword-planner/. Google Ads.
 - 
Google. 2025. Keyword Planner - Google Ads. https://ads.google.com/home/tools/ keyword-planner/. Accessed: 2025-05-18.
 - 
Ching-Lai Hwang and Kwangsun Yoon. 1981. Multiple Attribute Decision Making: Methods and Applications. Springer-Verlag, Berlin, Heidelberg.
 - 
Amruta Joshi and Rajeev Motwani. 2006. Keyword generation for search engine advertising. In Sixth IEEE International Conference on Data Mining - Workshops (ICDMW’06), pages 490–496.
 - 
Mu-Chu Lee, Bin Gao, and Ruofei Zhang. 2018. Rare query expansion through generative adversarial networks in search advertising. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD ‘18, page 500–508, New York, NY, USA. Association for Computing Machinery.
 - 
Yijiang Lian, Zhijie Chen, Jinlong Hu, Kefeng Zhang, Chunwei Yan, Muchenxuan Tong, Wenying Han, Hanju Guan, Ying Li, Ying Cao, Yang Yu, Zhigang Li, Xiaochun Liu, and Yue Wang. 2019. An end-toend generative retrieval method for sponsored search engine –decoding efficiently into a closed target domain. Preprint, arXiv:1902.00592.
 - 
Chin-Yew Lin. 2004. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74–81, Barcelona, Spain. Association for Computational Linguistics.
 - 
Microsoft. 2025. Keyword Planner - Microsoft Advertising. https://about.ads.microsoft.com/ja/ tools/planning/keyword-planner. Accessed: 2025-05-18.
 - 
Masato Mita, Soichiro Murakami, Akihiko Kato, and Peinan Zhang. 2024. Striking gold in advertising: Standardization and exploration of ad text generation. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 955–972, Bangkok, Thailand. Association for Computational Linguistics.
 - 
Han Nie, Yanwu Yang, and Daniel Zeng. 2019. Keyword generation for sponsored search advertising: Balancing coverage and relevance. IEEE Intelligent Systems, 34(5):14–24.
 - 
Shakked Noy and Whitney Zhang. 2023. Experimental evidence on the productivity effects of generative artificial intelligence. Science (New York, N.Y.), 381:187–192.
 - 
OpenAI. 2024. Gpt-4o system card. Preprint, arXiv:2410.21276.
 - 
Yang Wang, Zheyi Sha, Kunhai Lin, Chaobing Feng, Kunhong Zhu, Lipeng Wang, Xuewu Jiao, Fei Huang, Chao Ye, Dengwu He, Zhi Guo, Shuanglong Li, and Lin Liu. 2024. One-step reach: Llm-based keyword generation for sponsored search advertising. In Companion Proceedings of the ACM Web Conference 2024, WWW ‘24, page 1604–1608, New York, NY, USA. Association for Computing Machinery.
 - 
Zhao Wang, Briti Gangopadhyay, Mengjie Zhao, and Shingo Takamatsu. 2025a. OKG: On-the-fly keyword generation in sponsored search advertising. In Proceedings of the 31st International Conference on Computational Linguistics: Industry Track, pages 115–127, Abu Dhabi, UAE. Association for Computational Linguistics.
 - 
Zhao Wang, Sota Moriyama, Wei-Yao Wang, Briti Gangopadhyay, and Shingo Takamatsu. 2025b. Talk structurally, act hierarchically: A collaborative framework for llm multi-agent systems. arXiv preprint arXiv:2502.11098.
 - 
Hao Wu, Guang Qiu, Xiaofei He, Yuan Shi, Mingcheng Qu, Jing Shen, Jiajun Bu, and Chun Chen. 2009. Advertising keyword generation using active learning. In Proceedings of the 18th International Conference on World Wide Web, WWW ‘09, page 1095–1096, New York, NY, USA. Association for Computing Machinery.
 - 
Frank Xing. 2025. Designing heterogeneous llm agents for financial sentiment analysis. ACM Transactions on Management Information Systems, 16(1):1–24.
 - 
Shinn Yao et al. 2023. React: Synergizing reasoning and acting in language models. arXiv preprint arXiv:2210.03629.
 - 
Edmundas Kazimieras Zavadskas, Algimantas Zakarevicius, and Jurgita Antucheviciene. 2006. Evaluation of ranking accuracy in multi-criteria decisions. Informatica, 17(4):601–618.
 - 
Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, and Yoav Artzi. 2020. Bertscore: Evaluating text generation with bert. Preprint, arXiv:1904.09675.
 - 
Xiaoxi Zhong et al. 2023. Agentverse: Facilitating multi-agent collaboration and exploration with llms. arXiv preprint arXiv:2309.07864.
 
A Additional Experimental Explanation
Section titled “A Additional Experimental Explanation”A.1 Experiment Setting
Section titled “A.1 Experiment Setting”The experiment is run 5 times with different random seeds, which means we sample a different number of keywords for the multi-shot analysis in the benchmark results. The random seeds used are [42, 123, 456, 891, 777]. The temperature for the GPT-4o backbone is set as 0. The embedding model that is used to create the cluster is bert-base-cased. The API version of the used OpenAI models is based on the most recent available API at the time of writing.
The keyword generator and intention analysis is a GPT-4o model. The evaluator that is used to analyze the keyword quality is an o3-mini-high model.
For non-agentic models that are not equipped with tools, we first run the OMS the retrieve information and save this information as the product information. This product information will be used by other models to serve as the product information.
We randomly sample the 800 keywords as the evaluation data, and the left 200 keywords with their performance are used for the OKG Many-shot or the OMS.
Based on the log status from the Langsmith output, each generation took around 0.2-0.5$ based on the number of iterations.
A.1.1 A/B Test Setting
Section titled “A.1.1 A/B Test Setting”For the product for which we conducted an actual campaign to perform the A/B test, they are both deployed in the United States. The campaign lasted for 30 days from 2025 March 20, 2025, to April 20. The daily budget available for both systems is set at 4,000 yen. For every generation round, the newly generated keywords are deployed on Tuesday and Friday. This means the performance observation is collected every 4 or 3 days. We read the accumulated performance table from the API provided by the SSA platform. The newly generated keywords will be used to replace keywords that did not receive performance or whose performance did not meet our standards, as Google Ads only allows 50 keywords to be deployed at the same time.
Tool Signatures.
Section titled “Tool Signatures.”Search : Σ∗ → I, Generate : I → 2 K, RejectReflection : K → {0, 1}, RepeatedFilter : K → {0, 1}, SearchVolume : K → N, LexicalAnalysis : 2K → 2 P , CategoryReject : (2K, R) → 2 C .
A.2 Tools Introduction
Section titled “A.2 Tools Introduction”Reject Reflextion This tool will load keywords with poor performance and analyze why those keywords are bad. The Keyword Generator receives the analysis result to help it generate better keywords.
Search Tool This tool is based on SerpAPI 7 which provides Google search results for a query. For a given product, the product information is made by manually crafted information (if given) and retrieved information through automated searching. The keyword generator will first formalize a query about the product itself. Then, the keyword generator will decide whether the information retrieved is enough. If not, the Keyword Generator will formulate several queries related to the product and retrieve related information. This process is executed until the Keyword Generator thinks it has enough information for the keyword generation.
Algorithm 1 Iterative Keyword Generation with Constraint Feedback
Section titled “Algorithm 1 Iterative Keyword Generation with Constraint Feedback”Input: initial product query q0, maximum iterations T, desired keyword budget B, rejectedkeyword set R, deployed history set H, search–volume threshold τ , category rejection thresholdθOutput: validated keyword set K⋆1: info ← SEARCH(q0) ▷ Search Tool retrieves initial product information2: V ← ∅ ▷ temporary store for low–volume keywords3: for t = 1 to T do4: if ISENOUGH(info) then5: K ← GENERATE(info) ▷ LLM drafts keyword set6: Kv ← ∅ ▷ validated keywords7: for all k ∈ K do8: if k ∈ R then ▷ Reject Reflection9: continue10: else if k ∈ H then ▷ Repeated Keyword Filter11: continue12: else if SEARCHVOLUME(k) < τ then ▷ Search Volume Validator13: V ← V ∪ {k}14: else15: Kv ← Kv ∪ {k}16: end if17: end for18: if |Kv| ≥ B then19: return K⋆ ← Kv20: end if21: P ← LEXICALANALYSIS(V) ▷ common prefixes/suffixes22: UPDATEGENERATOR(P) ▷ adapt generation constraints23: C ← CATEGORYREJECT(V, θ)24: UPDATECATEGORIES(C) ▷ drop high–reject clusters25: else26: q ← NEXTQUERY(info)27: info ← info ∪ SEARCH(q)28: end if29: end for30: return K⋆ ← Kv ▷ May be empty if budget unmetRejected Keyword Filter For the deployed keywords, some keywords are above our budget after deployment. For example, a keyword with a high click-per-cost is saved into a local list, and we do not want the keyword generator to generate those keywords. The generated keywords will be passed to the Rejected Keyword Filter tool. This tool will check whether the generated keywords contain a rejected keyword and flag keywords that do. The Keyword Generator is asked to regenerate this keyword until all keywords are non-rejected.
Repeated Keyword Filter As we have already provided and deployed keywords in the multi-shot, since the performance of those keywords is already validated, we do not want the Keyword Generator to generate those keywords again. Therefore, like the Rejected Keyword Filter, this tool will flag keywords that are in the deployed history keyword. The Keyword Generator is asked to regenerate those flagged repeated keywords.
Search Volume Check When we deploy a keyword, the SSA platform will perform a simple check to see if a keyword has been searched a certain number of times in the past several months. If the generated keyword does not meet this requirement, the SSA platform will automatically disable it. Therefore, in order to make all generated keywords valid keywords. We have to call the API provided by the SSA platform to check the search volume of those generated keywords.
Keyword Lexical Analysis For those rejected keywords with either low performance or identified with low search volume, we will analyze their common n-gram patterns and their common prefix and suffix. If the count of such prefixes and suffixes is above a certain threshold, we will append those lexical features to the generation rules of the keyword generator. This makes sure the LLM keyword generator will not get stuck in an endless loop of regenerating keywords in order to pass the search volume validator.
Category Analysis As the keyword generator generates keywords according to the categories analyzed by the Keyword Generator agent. However, it is possible that if the category itself is bad, e.g, the category is too niche, then all generated keywords following this category will not pass the final search volume check, no matter how many times of regeneration. To avoid this situation, we calculate the percentage of rejected keywords for a category across multiple times of generations, then we can identify those categories that have a high reject rate for their keywords. If such a category is identified, the whole category is rejected, and the LLM agent is asked to generate a whole new category.
B Example Workflow
Section titled “B Example Workflow”We present an example workflow of the Tool-Return Guided Generation workflow in the Figure 5.
C Annotator Description for Human Preference over Generated Keywords
Section titled “C Annotator Description for Human Preference over Generated Keywords”We asked annotators (they are recruited within the company organization, thus no payment is provided) to evaluate advertisement keywords generated by five anonymized systems (Systems 1 to 5). We have obtained their consent to use their evaluation scores. The evaluation was conducted based on the following six criteria. Each criterion is rated on a scale from 0 to 5, where a higher score indicates better performance (except for Redundancy, where lower is better).
- Coverage: How well the keyword set covers the product features described in the product information text.
 - Relevance: Whether the keyword set avoids including irrelevant or weakly related terms.
 - Specificity: How specifically the keywords reflect the core characteristics of the product.
 - Redundancy: Whether there are redundant or similar keywords that could be removed (a lower score indicates better quality).
 - Search User Behavior Alignment: To what extent the keywords align with the search behavior of users looking for similar products.
 - Overall Quality: The overall evaluation of the keyword set, considering all the criteria above.
 
The following are the examples we provided to the human preference evaluation task for annotators.
Example 1: Product Information: The handmade dorayaki from “Kyoto Marushin” features chunky sweet bean paste carefully cooked from Hokkaido-grown azuki beans, sandwiched in fluffy, specially made pancakes. Free from preservatives and artificial coloring, it offers a natural sweetness and the original flavor of the ingredients. Ideal for both gifts and personal enjoyment.
- Keywords with high specificity but low relevance: Azuki sweets, Azuki dessert
 - Keywords aligned with user search behavior but low in specificity: Kyoto travel sweets
 - Keywords with low coverage and relevance: Dorayaki recipe
 - Good keyword examples: Kyoto dorayaki, Kyoto wagashi (traditional sweets), Handmade dorayaki, Additive-free dorayaki
 

Figure 5: An example workflow for keyword generation.
Example 2: Product Information: The SHARP Plasmacluster 7000 air purifier supports rooms up to 13 tatami mats in size, effectively removing pollen, dust, and PM2.5 particles. Its proprietary Plasmacluster technology also suppresses airborne viruses and mold. Equipped with a quiet mode, it is ideal for nighttime use and bedrooms. The long-life filter is easy to maintain.
- Keywords aligned with user search behavior but low in coverage or specificity: Highperformance filter, Popular air purifier
 - Keywords with high specificity but low coverage: Plasmacluster
 - Criteria for a good keyword set: Covers key product features (suitable for 13-tatami rooms, pollen/PM2.5 removal, quiet mode, long-life filter), avoids redundancy, and aligns well with user search behavior.
 
C.1 Human Preference Alignment
Section titled “C.1 Human Preference Alignment”We have three annotators for the agreement between the annotators for the generated keywords. We anonymized three annotators with A1, A2, and A3. We show the inner-annotator agreement between annotators in the Table 8. The results show that all annotators reach similar judgments regarding the keyword quality, supporting the validity of the experimental results.
| Metric | A1-A2 | A1-A3 | A2-A3 | 
|---|---|---|---|
| Coverage | 0.607 | 0.796 | 0.626 | 
| Relevance | 0.604 | 0.447 | 0.620 | 
| Specificity | 0.650 | 0.551 | 0.623 | 
| Redundancy | 0.640 | 0.596 | 0.579 | 
| Search Behavior | 0.520 | 0.695 | 0.687 | 
| Overall Quality | 0.590 | 0.829 | 0.625 | 
| Average | 0.601 | 0.652 | 0.627 | 
Table 8: Inter-annotator Agreement (Kendall’s Tau Correlation)
C.2 More Generated Keywords Examples
Section titled “C.2 More Generated Keywords Examples”Example: Product Information:
Section titled “Example: Product Information:”Sony’s Prediction One is a no-code AutoML tool that allows users without machine learning expertise to build high-accuracy models in a few clicks. Developed by Sony Network Communications, it automates everything from algorithm selection to hyperparameter tuning. It supports business use cases like demand forecasting and churn prediction, and provides APIs for integration into existing systems. A GUI-based interface and team collaboration features enable easy adoption across industries, addressing the AI talent shortage.
- 
o3: AutoML tool, No-code machine learning, High-accuracy model automation, Demand forecasting software, Churn prevention analysis, Failure prediction, Prediction One pricing, Prediction One deployment, Prediction One enterprise plan, AutoML comparison, AutoML tool comparison, Alternative to DataRobot
 - 
OKG: No-code machine learning, AutoML tool, Sony Prediction One, demand forecasting software, sales efficiency AI, DX promotion tool, data analysis software, automated prediction model, API prediction integration, AI tool for small businesses, DX solution for small businesses, business automation software
 - 
OMS: No-code machine learning, AutoML tool, AI prediction model, demand forecasting system, DX promotion tool, inventory optimization AI, Sony Prediction One, Sony AutoML, Sony AI tool, AI services for businesses, cloud AI implementation, AI tool with support
 - 
GPT-4o: High-accuracy model auto-construction, Sony AutoML, evidence-based prediction machine learning, one-click machine learning, Prediction One AI system, multi-industry optimization AI, one-click AI development
 - 
GPT-4.1: Sony Prediction One ML system, efficiency improvement ML tool, code-free AI, Prediction One business AI service, explainable prediction AI tool, easy AI implementation, GUI-based ML development, cloud AI tool, machine learning service for enterprises, sales analysis optimization tool
 
Example: Product Information:
Section titled “Example: Product Information:”Sony’s Neural Network Console is a no-code development environment for designing and training neural networks using drag-and-drop. It integrates with a Python API for hybrid GUI + code development and features automatic architecture search to speed up model optimization. Users can visually compare past experiments and export models in ONNX format for deployment on Edge AI devices. It’s used in real Sony products like Aibo and Xperia Ear, and supports beginners through tutorials and corporate users with PoC services.
- o3: Neural network development software, Neural network design tool, Deep learning GUI, AI model auto generation, Drag-and-drop machine learning, AutoML hyperparameter tuning, Automatic model structure search, Automated machine learning tool, ONNX export tool, Edge AI model deployment, Mobile AI integration
 - OKG: Easy neural network development, programming-free machine learning, beginner AI tutorial, image classification learning tool, time series prediction introduction, AI PoC solution, DX support service, AI prototype creation, Edge AI model development, ONNX export method, AI smartphone app integration
 - OMS:Drag-and-drop AI, deep learning development environment, neural network training, Pythonintegrated AI tool, AI prototyping, AI PoC support, DX promotion service, ONNX export, edge AI device, smartphone AI integration
 - GPT-4o: No-code AI development, no-code AI tool, edge AI implementation, AI talent shortage solution, AI business proposal, visualized AI tool, Sony AI product tool, AI transformation support
 - GPT-4.1: Model structure auto-optimization AI tool, AI tool with implementation examples, Pythonintegrated AI development, free AI tool, AI deployable to edge devices, deep learning development for beginners, neural network GUI development, no-code AI development
 
Example: Product Information:
Section titled “Example: Product Information:”Sony’s cloud-based attendance management system “AKASHI” supports diverse work styles from telework to flex-time. It simplifies HR operations with intuitive drag-and-drop rule settings, automatic creation of legally required records, and integration with face recognition AI for secure time tracking. The system enables real-time management from anywhere and helps companies reduce overtime through alerts and automated payroll integration. Offered as a monthly subscription, it’s suitable for mid-sized companies aiming to digitize attendance management.
- 
o3: Attendance management system, Cloud-based attendance software, Telework attendance tracking, Flex-time attendance, 36-agreement compliance tool, Paid leave management system, HR compliance software, Attendance-payroll API integration, Automated attendance data, HR system integration
 - 
OKG: Cloud-based attendance management, attendance software for small businesses, attendance system with facial recognition, HR efficiency solution, labor management automation, overtime reduction tool, telework attendance management, remote work clock-in system, remote attendance cloud, work style reform attendance tool, labor risk prevention system, digital back office transformation
 - 
OMS: Cloud attendance management, attendance management system implementation, attendance management solution, HR and labor management, business automation system, paperless management, telework attendance management, flex time attendance, work style reform system, payroll system integration, expense reimbursement integration, API-compatible attendance management
 - 
GPT-4o: Comprehensive attendance management solution, clock-in/out system, automated management creation, zero initial cost attendance system, work style reform, customizable attendance system, all-in-one attendance system, remote work automatic management
 - 
GPT-4.1: Sony cloud-based attendance management system, automated attendance management system, attendance system with implementation examples, data visualization, AKASHI attendance management, productivity improvement attendance management, law-compliant auto-updating attendance management, fraud prevention attendance system, remote clock-in system, automatic labor condition detection system
 
Example: Product Information:
Section titled “Example: Product Information:”Sony’s next-generation BRAVIA series was developed under the concept of “experiencing the world beyond the screen.” By leveraging the latest panel technologies—Mini LED and QD-OLED—it delivers dazzling brightness and deep blacks even in bright living rooms. The BRAVIA 8 II adopts third-gen QD-OLED panels with superior peak brightness and wide color gamut, reproducing vibrant colors for HDR content. The BRAVIA 9 features dense local dimming for rich contrast. With “Auto HDR Tone Mapping” and ultra-low latency ( 8 ms), it offers smooth gaming. The BRAVIA CAM adjusts picture and sound based on user position, ideal for both family living rooms and work desks.
- o3: Sony QD OLED TV, Mini LED BRAVIA comparison, BRAVIA 9 high brightness, BRAVIA game mode low latency, Sony TV for eSports, Game Menu settings, Auto HDR tone mapping, BRAVIA deep blacks, BRAVIA wide color volume, BRAVIA CAM auto adjustment, Sony TV sound optimization
 - OKG: QD OLED TV, Mini LED high picture quality, HDR image optimization, low latency gaming TV, game menu supported TV, large screen TV for families, automatic picture adjustment BRAVIA CAM, living room optimized TV
 - OMS: Mini LED TV, QD OLED TV, high brightness HDR TV, low latency gaming TV, TV with crosshair function for gaming, BRAVIA for eSports, BRAVIA CAM auto adjustment, optimal picture quality BRAVIA, TV with automatic viewing distance adjustment, cinema-level black reproduction TV, vivid color TV
 - GPT-4o: Living room movie watching TV, eSports optimized TV, BRAVIA for family living room, BRAVIA game mode low latency, automatic adjustment TV, Sony TV latest model
 - GPT-4.1: Auto-optimized setting TV, multi-purpose TV, low latency TV, Sony BRAVIA TV, BRAVIA new series, Mini LED TV, QD OLED TV, high picture quality TV for gaming, home use TV, cinema experience TV
 
D Prompts used in Methodology
Section titled “D Prompts used in Methodology”D.1 LLM_Intent Prompt
Section titled “D.1 LLM_Intent Prompt”The following prompt is used to implement the function LLM_Intent(r, Cj ) defined in Equation 3 for intent analysis of each keyword cluster:
LLM_Intent Prompt
Section titled “LLM_Intent Prompt”Product name {product_name} and the users searched for the following keywords to reach the website related to this product.
The related product information is given in the following: {product_information}
You are given the following clusters of keywords and their clicks: Cluster N: Keyword 1: Click, Cost, Conversion, Impression. …
Analyze the above cluster and its keywords. Especially check the common features of those keywords and explain why the user searched for those keywords.
D.2 LLM_Rank Prompt Template
Section titled “D.2 LLM_Rank Prompt Template”This prompt is constructed automatically after computing TOPSIS scores for each keyword and cluster. It is used as input to the function LLM_Rank(Cj ) defined in Equation 6. The structure below is generated programmatically using a for-loop over clusters and their ranked keywords:
LLM_Rank Prompt Template
Section titled “LLM_Rank Prompt Template”You are given clusters of keywords with their corresponding TOPSIS scores and rankings.
Each cluster is listed with its average TOPSIS score, and keywords within each cluster are sorted by their individual scores.
Use this information to determine which clusters or keywords should be prioritized for expansion or refinement.
Cluster: {cluster_name} (Avg Score: {cluster_avg_score}) 1. {keyword_1} — Score: {score_1} 2.{keyword_2} — Score: {score_2} 3. {keyword_3} — Score: {score_3} ...Cluster: {cluster_name} (Avg Score: {cluster_avg_score}) 1. {keyword_1} — Score: {score_1} 2. {keyword_2} — Score: {score_2} …
…
Please analyze the clusters and their keyword rankings. Focus on the strongest-performing clusters and keywords for generation, and suggest improvements for weaker ones.
D.3 LLM_Reflect Prompt Template
Section titled “D.3 LLM_Reflect Prompt Template”This prompt is used to implement the function LLM_Reflect(ki) = LLM(ki , Kt , r) as defined in Equation 7. The prompt is dynamically constructed using the product description, current generated keywords, and historical evaluations:
LLM_Reflect Prompt Template
Section titled “LLM_Reflect Prompt Template”You are given the intermediate generated keyword result (formatted as a dictionary with two main keys: {‘Branded’} and {‘Non-Branded’}) and the product information. You should evaluate the coherence between each keyword and the product.
For each keyword, give a score from 1 to 5 based on how well it represents the product information. Also, provide a reason for the score and suggest whether the keyword should be kept or replaced.
Scoring guide: 1: The keyword does not represent the product at all. 2: The keyword poorly represents the product. 3: The keyword somewhat represents the product. 4: The keyword represents the product well. 5: The keyword perfectly represents the product.
A good keyword should capture key product features and not be overly generic. Consider both semantic relevance and user search behavior.
Provide your output in the following dictionary format: { “keyword”: “{keyword}”, “score”: {score}, “reason”: “{reason}”, “suggestion”: “{keep/replace}” }
The generated keywords to evaluate are: {generated_keywords}
The product information is: {product_information}
Your evaluation history is: {history_evaluation}
Only evaluate keywords not already included in the history.
D.4 LLM_Assign Prompt Template
Section titled “D.4 LLM_Assign Prompt Template”This prompt is used to implement the function LLM_Assign(ki , C top ki ) as defined in Equation 8. The prompt is constructed dynamically using the target keyword, the top-3 nearest clusters (based on embedding similarity), and the product information:
LLM_Assign Prompt Template
Section titled “LLM_Assign Prompt Template”You are given a keyword: {keyword_tobe_decided}
The following three clusters are the closest clusters to this keyword based on embedding similarity:
The product information is: {product_information}
Cluster 1: {cluster_1_keywords}
Cluster 2: {cluster_2_keywords}
Cluster 3: {cluster_3_keywords}
Based on the product information and the semantic intent of each cluster, decide whether the given keyword should be assigned to one of the above clusters or treated as a new cluster.
Please respond with exactly one of the following options: Cluster 1, Cluster 2, Cluster 3, or New Cluster.
D.5 LLM_Generate Prompt Template
Section titled “D.5 LLM_Generate Prompt Template”This static prompt is given to the agent at initialization to define the overall task. It instructs the agent to generate advertising keywords through multi-step reasoning based on product information, past failures, and user search behavior.
LLM_Generate Prompt Template
Section titled “LLM_Generate Prompt Template”You are tasked with generating advertising keywords for {product_name}.
Your keywords must reflect the product’s key features and align with how users typically search online. Avoid technical terms and duplicates from previous keywords.
Before generation, use tools such as google_search to gather product information and reject_reflection to analyze failed keywords.
Your final output must be a dictionary-like string with two keys: “Branded” and “Non-Branded”, each containing 10 high-quality keywords.