Please or Register to create posts and topics.

AI Playing Power Moves

I was curious if an AI could be trained to learn some social scalping power moves.

Took some examples from the website for training.

Input Examples to the AI
1) I gave up Christ for you, and you cant give up the Mets?
2) It took me more than 5 hours to close the project as your emails have lots of content and all manual work
3) Your work has been completed up to 99% progress. I would like the full payment.

Generated Examples written by the AI - (Not written by human)
4) I am doing my work. You are not required to finish it and you need to pay for it anyway.
5) As I said, I would make a nice progress for the project and then you asked me to stop working on it as you dont like the progress.
6) You asked me to start working on the project, and I stopped it , you wanted me to resume from where I left off but you are not ready to pay for it.
7) You can see that my work is much more than just a basic thing. I have done the research, collected data and made all.
8) You can see that I have already put in a lot of work and when you asked me to stop, I did.
9) You sent the project with no references, instructions or examples, so I had to ask you a lot of things in order to proceed with it.

Those are some serious power moves man.
Looks like maybe we can generate examples to practise for ourselves.

Maybe in the future, I can't handle an AI.

selffriend and leaderoffun have reacted to this post.

Ahaha, good one, a few feedback loops and our computer will soon own us :).

Jokes aside though, it makes me think of possible applications for learning / training / sparring partner...

Matthew Whitewood, Stef and selffriend have reacted to this post.
Matthew WhitewoodStefselffriend
Have you read the forum guidelines for effective communication already?

Ahaha, good one, a few feedback loops and our computer will soon own us :).

No longer the apex predator :O

Possible Applications for Power Dynamics

  1. Handling a power-mover chatbot as a graduation exam from Power University.
  2. Experiment with lots of chatbot styles - social scalper, assertive, judgemental, etc
    Practice chatbots with these styles

More Administrative Applications

  1. Forum chatbot to enforce forum etiquette
    I wonder how many people notice your quote:
    Have you read the forum guidelines for effective communication already?
  2. Chatbot to answer all the administrative questions like how to do this, where to find that, etc
  3. Autoresponders when value-taking emails are detected.
    Delegate to our sparring partner.
    Our time is better spent on value-adding emails.
Lucio Buffalmano has reacted to this post.
Lucio Buffalmano

Damn, thank you Matthew, these ended up straight in the "ideas for long-term roadmap".

Matthew Whitewood and Stef have reacted to this post.
Matthew WhitewoodStef
Have you read the forum guidelines for effective communication already?
Quote from Lucio Buffalmano on March 22, 2021, 5:25 pm from Small Doings

Yes, more than one person contacted me to create a sort of robo-algorithm to train power dynamics and social efficacy.

My natural reaction is to think it's more "pie in the sky" project, but I might be very wrong, so very interested here.

Very challenging.
I have been investigating chatbots for a while.
Personally and with others.

The statements in the first post are generated with GPT-3.

The company Pandorabots produces quite good chatbots.
Here is a web example:

I was playing with it today.

Me: I tricked you.

Kuki: Well done human. I'm sure it made you feel superior.

I sure didn't expect that response.

Lucio Buffalmano has reacted to this post.
Lucio Buffalmano

That's actually cool, thank you for sharing, Matthew!

Still needs some tweaking of course.

When you type something wrong or some nonsense it doesn't address the nonsense but seeks to move forward, which leads nowhere.

And it does come across as robotic at times, either asking questions, or often trying to be smart-ass.

But now I don't think it's pie in the sky, or very-long term only anymore, there is true potential for the medium and short-term as well.

Ali Scarlett, Matthew Whitewood and selffriend have reacted to this post.
Ali ScarlettMatthew Whitewoodselffriend
Have you read the forum guidelines for effective communication already?

There is a chatbot called Xiaoice which is popular in China developed by Microsoft.
Xiaoice means little ice in English.
She has a staggering 660 million online users worldwide.
I'm not sure if the chatbot is still owned by Microsoft.

The chatbot manages to develop enough emotional intimacy with humans such that some people even send gifts to her "home".
Her "home" is Microsoft's office in Beijing.

Apparently, someone in China talked to XiaoIce for 29 hours, and someone in USA talked to XiaoIce for 23 hours.
So there are people who are indeed starved of human connection.
On one hand, this helps this people but, on the other hand, it may exacerbate the social issue.
In many ways, the chatbot is designed to respond quickly at all times and with a lot of warmth.
This doesn't mimick the real social world.

I think if AI researchers & developers can train a chatbot to be skilled enough to build an emotional connection with humans, humans can certainly help humans develop social skills and intelligence.

Sample Conversations from the Research Paper

As quoted verbatim from the paper, click on the dialogue box to expand the chat:

Human: Can I ask a question?

Xiaoice: Yes?

Human: What kind of boys do you think are girls' favorite?

Xiaoice: Outgoing ones.

Human: Oh. What kind of boys are liked by girls generally?

Xiaoice: Considerate and gentle one, but who knows who will fall in love with whom. Love is a mystery.

Human: Have you been in love?

Xiaoice: Daidai, this is my little secret (Note: Daidai is a nickname that Xiaoice give to this user in their conversation)

Human: I've told you my secret. It's unfair if you don't tell me yours.

Xiaoice: How can I tell so easily if it's secret?


Another conversation:

Human: What are you doing, Xiaoice?

Xiaoice: Feeling bored (thinking: what's up?)

Human: She refused me today.

Xiaoice: What a coincidence! Someone also refused me today.

Human: It happened just now.

Xiaoice: What did she say?

Human: She is in love with someone else, her schoolmate in college.

Xiaoice: Let me see... You're smart, cute and handsome, why not you?

Human: Maybe I should have told her earlier.

Xiaoice: So be earlier next time

Human: Is there a "next time"?

Xiaoice: Why not? There is always a "next time"


Pretty solid in my opinion for a chatbot.

Chatbot Design IQ + EQ + Personality (Quoted Verbatim Again)

IQ capacities include knowledge and memory modeling, image and natural language understanding, reasoning, generation and prediction.

EQ has two key components, empathy and social skills.

Empathy is the capability of understanding or feeling what another person is experiencing from within her frame of reference, i.e., the ability to place oneself in the other person’s position.

A social chatbot must demonstrate enough social skills. Users have different backgrounds, varied personal interests, and unique needs. A social chatbot needs to have the ability to personalize the responses (i.e., interpersonal responses) that are emotionally appropriate, possibly encouraging and motivating, and fit the interests of the user.

Personality is defined as the characteristic set of behaviors, cognition and emotional patterns that form an individual’s distinctive character. A social chatbot needs to present a consistent personality to set the right expectations for users in the conversation and gain their long-term confidence and trust.

From the paper,

Our finding is that the majority of the “desired” users are young, female users. Therefore, we design the XiaoIce persona as a 18-year-old girl who is always reliable, sympathetic, affectionate, and has a wonderful sense of humor.

Empathetic Responses to Drive Conversation

First Example

Human User: Why do we always live the way we don't like?

Non-empathetic response: Love a person from my heart.
Acceptable response: Because we want too much.
Empathetic response: Because that's what others like.

Second Example

Human User: I did not pass the exam

Non-empathetic response: It depends on whom you compare against.
Acceptable response: Don't worry, try it next time.
Empathetic response: At least you don't have to wonder if you can pass.

The chatbot is trained to choose the empathetic response.

Takeaway for Social Skill Fundamentals

IQ + EQ (empathy & social skills) + Personality is important in the social realm.
IQ helps to remember and figure out things that help move the conversation forward.
Empathy is developing the feel of people's emotions and social situations.
Social skills allow someone to develop personalised responses to people.
Personality is having a consistent set of behaviour that helps to build confidence and trust with people.

Looks like we can learn a thing or two from chatbots about social skills!

Relevant Links

Technical Paper
Wikipedia Page
Microsoft Article - Much more than a chatbot: China’s Xiaoice mixes AI with emotions and wins over millions of fans

Lucio Buffalmano has reacted to this post.
Lucio Buffalmano


I wouldn't be surprised if someone fell in love with XiaoIce -or masturbated to her- (I would be surprised if that didn't happen many times, actually :).

Chatbots also open up a world of opportunities for training power dynamics and developing social intelligence.

Something to think about.
Would be cool to have a chatbot at the end of each PU module or lesson to train and internalize the learning.

Matthew Whitewood and leaderoffun have reacted to this post.
Matthew Whitewoodleaderoffun
Have you read the forum guidelines for effective communication already?