The post Want Better Results From an AI Chatbot? Be a Jerk appeared on BitcoinEthereumNews.com. In brief Penn State researchers found that “very rude” prompts outperformed “very polite” ones in accuracy. The results contradict prior studies claiming LLMs respond better to a courteous tone. Findings imply that tone itself, once dismissed as etiquette, may be a hidden variable in prompt engineering. Being polite might make you a better person, but it could make your AI assistant a dumbass. A new Penn State study finds that impolite prompts consistently outperform polite ones when querying large language models such as ChatGPT. The paper, “Mind Your Tone: Investigating How Prompt Politeness Affects LLM Accuracy,” reports that “very rude” prompts produced correct answers 84.8% of the time, compared with 80.8% for “very polite” ones. That’s a small but statistically significant reversal of previous findings, which suggested models mirrored human social norms and rewarded civility. “Contrary to expectations,” wrote authors Om Dobariya and Akhil Kumar, “impolite prompts consistently outperformed polite ones… suggesting that newer LLMs may respond differently to tonal variation.”  The conflicting science of prompt engineering The findings reverse expectations from a 2024 study, “Should We Respect LLMs? A Cross-Lingual Study on the Influence of Prompt Politeness on LLM Performance,” which found that impolite prompts often degraded model performance, while excessive politeness offered no clear benefit. That paper treated tone as a subtle but mostly stabilizing influence. The new Penn State results flip that narrative, showing that—at least for ChatGPT-4o—rudeness can sharpen accuracy, suggesting that newer models no longer behave as social mirrors but as strictly functional machines that prize directness over decorum. However, they support more recent research from the Wharton School into the emerging craft of prompt engineering—phrasing questions to coax better results from AIs. Tone, long treated as irrelevant, increasingly appears to matter almost as much as word choice. The researchers rewrote 50 base questions… The post Want Better Results From an AI Chatbot? Be a Jerk appeared on BitcoinEthereumNews.com. In brief Penn State researchers found that “very rude” prompts outperformed “very polite” ones in accuracy. The results contradict prior studies claiming LLMs respond better to a courteous tone. Findings imply that tone itself, once dismissed as etiquette, may be a hidden variable in prompt engineering. Being polite might make you a better person, but it could make your AI assistant a dumbass. A new Penn State study finds that impolite prompts consistently outperform polite ones when querying large language models such as ChatGPT. The paper, “Mind Your Tone: Investigating How Prompt Politeness Affects LLM Accuracy,” reports that “very rude” prompts produced correct answers 84.8% of the time, compared with 80.8% for “very polite” ones. That’s a small but statistically significant reversal of previous findings, which suggested models mirrored human social norms and rewarded civility. “Contrary to expectations,” wrote authors Om Dobariya and Akhil Kumar, “impolite prompts consistently outperformed polite ones… suggesting that newer LLMs may respond differently to tonal variation.”  The conflicting science of prompt engineering The findings reverse expectations from a 2024 study, “Should We Respect LLMs? A Cross-Lingual Study on the Influence of Prompt Politeness on LLM Performance,” which found that impolite prompts often degraded model performance, while excessive politeness offered no clear benefit. That paper treated tone as a subtle but mostly stabilizing influence. The new Penn State results flip that narrative, showing that—at least for ChatGPT-4o—rudeness can sharpen accuracy, suggesting that newer models no longer behave as social mirrors but as strictly functional machines that prize directness over decorum. However, they support more recent research from the Wharton School into the emerging craft of prompt engineering—phrasing questions to coax better results from AIs. Tone, long treated as irrelevant, increasingly appears to matter almost as much as word choice. The researchers rewrote 50 base questions…

Want Better Results From an AI Chatbot? Be a Jerk

For feedback or concerns regarding this content, please contact us at [email protected]

In brief

  • Penn State researchers found that “very rude” prompts outperformed “very polite” ones in accuracy.
  • The results contradict prior studies claiming LLMs respond better to a courteous tone.
  • Findings imply that tone itself, once dismissed as etiquette, may be a hidden variable in prompt engineering.

Being polite might make you a better person, but it could make your AI assistant a dumbass.

A new Penn State study finds that impolite prompts consistently outperform polite ones when querying large language models such as ChatGPT. The paper, “Mind Your Tone: Investigating How Prompt Politeness Affects LLM Accuracy,” reports that “very rude” prompts produced correct answers 84.8% of the time, compared with 80.8% for “very polite” ones.

That’s a small but statistically significant reversal of previous findings, which suggested models mirrored human social norms and rewarded civility.

“Contrary to expectations,” wrote authors Om Dobariya and Akhil Kumar, “impolite prompts consistently outperformed polite ones… suggesting that newer LLMs may respond differently to tonal variation.”

The conflicting science of prompt engineering

The findings reverse expectations from a 2024 study, “Should We Respect LLMs? A Cross-Lingual Study on the Influence of Prompt Politeness on LLM Performance,” which found that impolite prompts often degraded model performance, while excessive politeness offered no clear benefit.

That paper treated tone as a subtle but mostly stabilizing influence. The new Penn State results flip that narrative, showing that—at least for ChatGPT-4o—rudeness can sharpen accuracy, suggesting that newer models no longer behave as social mirrors but as strictly functional machines that prize directness over decorum.

However, they support more recent research from the Wharton School into the emerging craft of prompt engineering—phrasing questions to coax better results from AIs. Tone, long treated as irrelevant, increasingly appears to matter almost as much as word choice.

The researchers rewrote 50 base questions in subjects such as math, science, and history across five tonal levels, from “very polite” to “very rude,” yielding 250 total prompts. ChatGPT-4o was then asked to answer each, and its responses were scored for accuracy.

The implications stretch beyond etiquette. If politeness skews model accuracy, then it calls into question claims of objectivity in AI outputs. Rude users might, paradoxically, be rewarded with sharper performance.

Machine logic and human norms clash

Why might blunt or rude phrasing boost accuracy? One theory: polite prompts often include indirect phrasing (“Could you please tell me…”), which may introduce ambiguity. A curt “Tell me the answer” strips away linguistic padding, giving models clearer intent.

Still, the findings underscore how far AI remains from human empathy: the same words that smooth social exchange between people might muddy machine logic.

The paper hasn’t yet been peer-reviewed, but it’s already generating buzz among prompt engineers and researchers, who see it as a sign that future models may need social calibration—not just technical fine-tuning.

Regardless, it’s not like this should come as a shock to anyone. After all, OpenAI CEO Sam Altman did warn us that saying please and thank you to ChatGPT was a waste of time and money.

Generally Intelligent Newsletter

A weekly AI journey narrated by Gen, a generative AI model.

Source: https://decrypt.co/344059/want-better-results-from-ai-chatbot-be-jerk

Market Opportunity
null Logo
null Price(null)
--
----
USD
null (null) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact [email protected] for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.

You May Also Like

Solana Price Prediction: Litecoin Latest Updates As Pepeto Gains Buzz With Analysts Calling 100x Potential

Solana Price Prediction: Litecoin Latest Updates As Pepeto Gains Buzz With Analysts Calling 100x Potential

The post Solana Price Prediction: Litecoin Latest Updates As Pepeto Gains Buzz With Analysts Calling 100x Potential appeared first on Coinpedia Fintech News The discussion around Solana price prediction and Litecoin price prediction is shifting toward a different headline: Pepeto (PEPETO). While majors like Solana and Litecoin still draw eyes, momentum is tilting to Pepeto, an Ethereum memecoin with working utility. The project has already raised more than $6.6 million in presale with entry at $0.000000153. Analysts and …
Share
CoinPedia2025/09/18 12:42
Taiko and Chainlink to Unleash Reliable Onchain Data for DeFi Ecosystem

Taiko and Chainlink to Unleash Reliable Onchain Data for DeFi Ecosystem

Taiko and Chainlink Data Streams to deliver secure, high-speed onchain data by empowering next-generation DeFi protocols and institutional-grade adoption.
Share
Blockchainreporter2025/09/18 06:10
DOGE ETF Hype Fades as Whales Sell and Traders Await Decline

DOGE ETF Hype Fades as Whales Sell and Traders Await Decline

The post DOGE ETF Hype Fades as Whales Sell and Traders Await Decline appeared on BitcoinEthereumNews.com. Leading meme coin Dogecoin (DOGE) has struggled to gain momentum despite excitement surrounding the anticipated launch of a US-listed Dogecoin ETF this week. On-chain data reveals a decline in whale participation and a general uptick in coin selloffs across exchanges, hinting at the possibility of a deeper price pullback in the coming days. Sponsored Sponsored DOGE Faces Decline as Whales Hold Back, Traders Sell The market is anticipating the launch of Rex-Osprey’s Dogecoin ETF (DOJE) tomorrow, which is expected to give traditional investors direct exposure to Dogecoin’s price movements.  However, DOGE’s price performance has remained muted ahead of the milestone, signaling a lack of enthusiasm from traders. According to on-chain analytics platform Nansen, whale accumulation has slowed notably over the past week. Large investors, with wallets containing DOGE coins worth more than $1 million, appear unconvinced by the ETF narrative and have reduced their holdings by over 4% in the past week.  For token TA and market updates: Want more token insights like this? Sign up for Editor Harsh Notariya’s Daily Crypto Newsletter here. Dogecoin Whale Activity. Source: Nansen When large holders reduce their accumulation, it signals a bearish shift in market sentiment. This reduced DOGE demand from significant players can lead to decreased buying pressure, potentially resulting in price stagnation or declines in the near term. Sponsored Sponsored Furthermore, DOGE’s exchange reserve has risen steadily in the past week, suggesting that more traders are transferring DOGE to exchanges with the intent to sell. As of this writing, the altcoin’s exchange balance sits at 28 billion DOGE, climbing by 12% in the past seven days. DOGE Balance on Exchanges. Source: Glassnode A rising exchange balance indicates that holders are moving their assets to trading platforms to sell rather than to hold. This influx of coins onto exchanges increases the available supply in…
Share
BitcoinEthereumNews2025/09/18 05:07