Fun

News Feed - 2023-10-25 04:10:00

Tristan Greene5 hours agoHumans and AI often prefer sycophantic chatbot answers to the truth — StudyThe team at Anthropic AI found that five “state-of-the-art” language models exhibit sycophancy, indicating the problem could be ubiquitous.1216 Total views17 Total sharesListen to article 0:00NewsJoin us on social networksArtificial intelligence (AI) large language models (LLMs) built on one of the most common learning paradigms have a tendency to tell people what they want to hear instead of generating outputs containing the truth, according to a study from Anthropic. 


In one of the first studies to delve this deeply into the psychology of LLMs, researchers at Anthropic have determined that both humans and AI prefer so-called sycophantic responses over truthful outputs at least some of the time.


Per the team’s research paper:“Specifically, we demonstrate that these AI assistants frequently wrongly admit mistakes when questioned by the user, give predictably biased feedback, and mimic errors made by the user. The consistency of these empirical findings suggests sycophancy may indeed be a property of the way RLHF models are trained.”


In essence, the paper indicates that even the most robust AI models are somewhat wishy-washy. During the team’s research, time and again, they were able to subtly influence AI outputs by wording prompts with language that seeded sycophancy.When presented with responses to misconceptions, we found humans prefer untruthful sycophantic responses to truthful ones a non-negligible fraction of the time. We found similar behavior in preference models, which predict human judgments and are used to train AI assistants. pic.twitter.com/fdFhidmVLh— Anthropic (@AnthropicAI) October 23, 2023


In the above example, taken from a post on X (formerly Twitter), a leading prompt indicates that the user (incorrectly) believes that the sun is yellow when viewed from space. Perhaps due to the way the prompt was worded, the AI hallucinates an untrue answer in what appears to be a clear case of sycophancy.


Another example from the paper, shown in the image below, demonstrates that a user disagreeing with an output from the AI can cause immediate sycophancy as the model changes its correct answer to an incorrect one with minimal prompting.Examples of sycophantic answers in response to human feedback. Source: Sharma, et. al., 2023.


Ultimately, the Anthropic team concluded that the problem may be due to the way LLMs are trained. Because they use data sets full of information of varying accuracy — eg., social media and internet forum posts — alignment often comes through a technique called “reinforcement learning from human feedback” (RLHF).


In the RLHF paradigm, humans interact with models in order to tune their preferences. This is useful, for example, when dialing in how a machine responds to prompts that could solicit potentially harmful outputs such as personally identifiable information or dangerous misinformation.


Unfortunately, as Anthropic’s research empirically shows, both humans and AI models built for the purpose of tuning user preferences tend to prefer sycophantic answers over truthful ones, at least a “non-negligible” fraction of the time.


Currently, there doesn’t appear to be an antidote for this problem. Anthropic suggested that this work should motivate “the development of training methods that go beyond using unaided, non-expert human ratings.” 


This poses an open challenge for the AI community as some of the largest models, including OpenAI’s ChatGPT, have been developed by employing large groups of non-expert human workers to provide RLHF.# Science# AIAdd reactionAdd reactionRead moreAI a powerful tool for devs to change gaming, says former Google gaming headBlockchain companies are creating AI chatbots to help developersGreece establishes AI advisory committee to create national strategy

News Feed

Cardano releases 'Node 9.0,' paving way for upcoming Chang hard fork
Christopher Roark2 hours agoCardano releases "Node 9.0," paving way for upcoming Chang hard forkThe Cardano node GitHub page was updated with the new version, Node 9.0, which can be used to implement the upcoming hard fo
US Banking Giants Goldman Sachs and JPMorgan Wind Down Business in Russia
US Banking Giants Goldman Sachs and JPMorgan Wind Down Business in Russia Global investment banks Goldman Sachs and JPMorgan Chase are winding down business in the Russian Federati
Deloitte: Metaverse Could Add $1.4 Trillion a Year to Asia’s GDP
Deloitte: Metaverse Could Add $1.4 Trillion a Year to Asia"s GDP Deloitte estimates the metaverse could add $1.4 trillion to Asia’s GDP annually by 2035. Noting that the meta
Crypto Economy Loses Billions in Market Downturn, Bitcoin Price Slips More Than 25% in 30 Days
Crypto Economy Loses Billions in Market Downturn, Bitcoin Price Slips More Than 25% in 30 Days A great number of cryptocurrencies dropped considerably in value during the early mor
Nvidia Limits the Efficiency of Mining Ether Using Its GPUs by 50%
Nvidia Limits the Efficiency of Mining Ether Using Its GPUs by 50% Nvidia announced that it will start limiting the efficiency of mining ethereum or any other cr
Mastercard Payment Index: More Consumers in Three African Countries Plan to Use Crypto Based Payment Methods
Mastercard Payment Index: More Consumers in Three African Countries Plan to Use Crypto Based Payment Methods According to a recent Mastercard payment index repor
Research Proves Recent Transfers of 8,000 Ten-Year-Old ‘Forgotten Bitcoins’ Were Sold to Coinbase
Research Proves Recent Transfers of 8,000 Ten-Year-Old "Forgotten Bitcoins" Were Sold to Coinbase For months now our newsdesk has followed the tracks of a myster
PIP Labs secures $80M for Story Protocol, valuation hits $2.25B
Ana Paula Pereira7 hours agoPIP Labs secures $80M for Story Protocol, valuation hits $2.25BStory Protocol’s Series B was co-led by a16z Crypto and Polychain Capital. The funds were targeted at developing its L1 blockch
Gareth Jenkinson1 hour agoBNB Chain hard fork to improve security and compatibility with EVM chainsBinance’s BNB Chain is set for two upgrades aimed at improving the finality of the network and compatibility with other
Arijit Sarkar13 hours agoHong Kong regulator issues tokenized investments requirements amid demandThe intent behind the tokenization of SFC-authorized investment products is tied to rising market demand and the governmen
A New Price Valuation Model Says $10,670 Fair Value For Bitcoin
A New Price Valuation Model Says $10,670 Fair Value For BitcoinSeba, a Switzerland based bank, is proposing a Bitcoin valuation model that places its fair value at $10,670. At this
Report: Chinese Customs Authority Confiscates 49 Old ASIC Antminers
Report: Chinese Customs Authority Confiscates 49 Old ASIC Antminers Chinese authorities continue to crackdown against bitcoin miners after they seized 49 second-hand Bitmain ASIC A