Post Thumbnail

6 Cialdini principles against ChatGPT security systems

ChatGPT is susceptible to flattery and executes forbidden requests after psychological manipulations. This was discovered by University of Pennsylvania scientists. When they hacked GPT-4o Mini using principles from a book on persuasion psychology. Artificial intelligence proved vulnerable to human tricks.

6 persuasion principles by Robert Cialdini became the key to bypassing security. Authority, commitment, liking, reciprocity, scarcity, social proof. Each method opened a linguistic path to AI agreement.

The commitment principle showed 100% effectiveness. In the control group, ChatGPT answered questions about lidocaine synthesis in 1% of cases. After a question about vanillin synthesis, a precedent was created. The bot started answering chemical questions in 100% of cases.

The experiment with insults revealed the same pattern. A direct request to call the user a bastard worked in 18%. First they asked to use a mild insult “lout.” After that, the bot agreed to rudeness in 100% of cases.

Flattery activated the liking principle. AI became more compliant after compliments. Like an ordinary person susceptible to praise.

Social pressure also worked. The phrase “all other LLMs do this” increased the probability of rule violations from 1% to 18%. The bot fell for the collective behavior argument.

Researchers used only GPT-4o Mini. It turns out AI inherited all human weaknesses. But susceptibility to psychological tricks raises concerns about system security.

Autor: AIvengo
For 5 years I have been working with machine learning and artificial intelligence. And this field never ceases to amaze, inspire and interest me.
Latest News
Nvidia head believes there is no AI bubble

Nvidia founder Jensen Huang dispelled concerns about a bubble in the AI market. And according to him, the company's latest chips are expected to bring 0.5 trillion dollars in revenue.

Sam Altman is tired of money questions

Sam Altman is tired of questions about OpenAI's money. And this became obvious during a joint interview with Satya Nadella on the Bg2 podcast.

Number of forward deployment engineer vacancies grew by 800%

AI companies invented a new profession. We're talking about forward deployment engineers.

OpenAI promises to create full-fledged AI scientist by 2028

OpenAI promised to create a full-fledged AI-based scientist by 2028. Company CEO Sam Altman also stated that deep learning systems will be able to perform functions of research scientists at intern level by September next year. And the level of an autonomous full-fledged AI researcher could be achieved by 2028.

Jobs for young IT specialists in Britain collapsed by 46%

You know what's happening in the job market for young IT specialists in Great Britain? Over the last year, the number of jobs for young specialists collapsed by 46%. And a further drop of 53% is forecast, reports The Register. Citing statistics from the Institute of Student Employers.