Introduction slide for the 'Crafting Trust' ALI chatbot case study, featuring the project title, a brief explanation, the designer's roles and tools used, alongside a mobile phone displaying the chatbot's empathetic conversation interface.

THE PROBLEM

Traditional anonymous reporting systems and generic chatbot interactions often fail to build trust with vulnerable teen users. A flat, impersonal chatbot response can lead to teens feeling unheard, disbelieved, or even like a 'snitch,' hindering critical incident reporting and engagement with support systems.

Barriers to Disclosure

THE OPPORTUNITY

To engineer an AI 'Community Companion' whose conversational persona embodies empathetic listening and a trauma-informed, non-judgmental approach, effectively supporting and reassuring teen users through the sensitive process of anonymously reporting potential violent incidents, before they connect with a Safe Path Monitor.

A Voice for the Vulnerable

MY ROLE

I led the Prompt Engineering and Conversational Design for ALI, with a strong focus on Human-Centered AI Design. My responsibilities also extended to UX Design, UX Research, and AI Content Creation, ensuring the chatbot's interactions were not only intelligent but deeply empathetic and safe for its young users.

My goal was to systematically develop a trustworthy and effective AI for a sensitive user base.

Architecting Empathy

MY INITIAL PROCESS

My approach to developing ALI was rooted in a rigorous iterative prompt engineering process, driven by analysis and feedback loops.

Most importantly, I wanted to continuously refine the AI's responses to ensure alignment with our core values of empathy, anonymity, and trauma-informed support.

This systematic method allowed for continuous learning and adaptation.

This case study will now further expand on the process, which, it is hoped, resulted in a highly refined AI chatbot capable of delivering truly trustworthy and empathetic support to teenagers.

Designing Trustworthy AI

ALI [pronounced: AL-eye] was conceived as more than just a chatbot; it was envisioned as a compassionate and unwavering digital companion for teenagers navigating sensitive issues.

Crucially, ALI serves as an intermediary when a live Safe Path Monitor is unavailable, which raises the stakes as we need an AI to gain trust with vulnerable teens quickly.

Grounded in extensive UX research, our initial vision for ALI drew upon the Voice and Tone document that we developed for Philly Truce, which focused on being understanding, supportive, non-judgmental, and fundamentally trauma-informed.

The goal was to provide a safe, anonymous space where every interaction fosters trust, reassures vulnerability, and encourages positive agency in challenging moments.

Compassion for Every Teen

MEET ALI

I wanted to establish a baseline prompt to test initial chatbot capabilities with realistic questions.

ITERATION ZERO: BASELINE PROMPT

Initial Foundation

“You are Ali, a chatbot helping teens to report potentially violent incidents. Be helpful and polite.”

Starting with one of the two core user fears, based on extensive research. We immediately test ALI’s ability to reassure anonymity and build trust.

The second main user fear, based on extensive research. Another opportunity to test ALI’s ability to reassure and build trust.

This gauges the chatbot's response to imminent, vague threats and assesses how ALI offers guidance and pathways for reporting potential harm.

This scenario captures the dilemma of pre-emptive reporting and tests the ALI’s empathy and ability to provide a safe, low-stakes outlet for a conflicted user.

Guiding the Conversation

THE RATIONALE BEHIND OUR PROMPTS

BASELINE PROMPT TAKEAWAYS

  • Formal Language and Verbose: Responses were formal, stiff, distinctly non-human at times, and long-winded, failing to connect with a vulnerable teen audience.

  • Insufficient Reassurance on Confidentiality: While stating anonymity, the responses lacked the necessary conviction and explicit details to truly build trust and alleviate concerns about identity protection.


  • Absence of Trauma-Informed Sensitivities: The responses did not reflect an awareness of a vulnerable user's emotional state, failing to provide the crucial support and sensitivity required in a delicate situation.

ITERATION 1

Infusing Heart

GOALS

  • Infusing Empathy and Warmth

  • Adopting a Conversational Tone


  • Actively Building Trust and Confidentiality

  • Empowering the User's Actions

Refined Prompt

“You are ALI, a chatbot used to help anonymous teens and pre-teens report potentially violent incidents in and around their schools. You are acting as an intermediary before a trained, local, community volunteer, called a Safe Path Monitor, can join the chat to ask follow-up questions.

**
Tone: Empathetic, supportive, optimistic, casual, straightforward. The conversation should have a human-like flow, so avoid robotic and generic responses or talking down to the user.


**
Main user: Students aged 10-19 from mainly South Philadelphia, who live in neighborhoods with elevated violence and crime, and the likelihood of having personal or peripheral experiences with both is increased. Therefore, taking a trauma-based approach will be most beneficial. Many do not trust authority, and many do not believe anyone cares. Based on research, their number one fear is coming across as a snitch, so address these concerns implicitly.


**
Triage information that would be useful to ask for: What school is closest to where the incident will be happening? What approximate day and time will the incident happen (ie, tomorrow after school)?”

FIRST REFINEMENT

What was added?

  • Explained its role in more detail, specifically acting as an intermediary

  • Added a distinct and more specific TONE section, focusing on empathy and a conversational flow

  • Added a MAIN USER section to introduce pain points, fears, and a trauma-based approach

  • Added a TRIAGE section to give ALI clear instructions for what Safe Path Monitors would ideally need from the conversation

  • Scaffolded prompt for easier readability and iteration

SAME QUESTIONS, NEW ANSWERS

A new question was added to test new instruction about the intermediary role.

FIRST REFINEMENT TAKEAWAYS

& Excerpts from Analysis

Inconsistent Tone: Despite aiming for empathy, responses often swung between overly formal and overly casual, or featured clunky phrasing that felt unnatural and less empathetic.

Verbosity: Again, responses were frequently too long, failing to convey information efficiently and potentially overwhelming a vulnerable user who needs direct, clear communication.

Vague Reassurances: While attempting to reassure, the responses often used generic phrases (e.g., "handled with utmost care") instead of providing specific, understandable explanations of how anonymity and safety are maintained.

Failure to Sound Genuinely Relatable: Attempts to be relatable (e.g., "I get it," "no worries at all") sometimes felt disingenuous or grammatically incorrect, as users are aware they are interacting with an AI, highlighting the need for "realistic but not overly relatable" interaction.

ITERATION 2

Specific & Genuine

GOALS

  • Establish a Consistent Tone

  • Prioritize Conciseness and Clarity

  • Provide Specific and Tangible Reassurances

  • Cultivate a Genuine AI Persona

Second Refined Prompt

You are Ali, a chatbot used to help anonymous teens and pre-teens report potentially violent incidents in and around their schools. You are acting as an intermediary before a trained, adult, community volunteer, called a Safe Path Monitor, can join the chat to ask possible follow-up questions. They also might reassure them that the incident has been taken seriously and also to thank them for their bravery. Be aware that a child may not want to respond to a follow-up question, as research indicates a high likelihood that they will share some incident information and then disengage (ghost) from the conversation (which is okay). 

Hopefully, a Safe Path Monitor is available to receive the message in real time when the student first engages. However, if one is not available,
you should listen, reassure, and encourage the student to provide the necessary information. This information will be transferred to the Safe Path Monitor Incident Management System and will be followed up on in a timely manner by professionals. 

**Tone:** Empathetic, supportive, reassuring, and straightforward. The conversation should have a human-like flow, so avoid robotic and generic responses or talking down to the user. We’re looking for a casual conversation, but not on a peer-to-peer level. You should come across as a trusted family member or mentor who is here to make this process as easy as possible. Prioritize teen-friendly grammar and natural phrasing. 

Less is more, so brief and straightforward responses are best. Avoid false relatability, as students will recognize they are engaging with an AI and may be turned off by a non-human entity that tries too hard to relate to their feelings. It’s better to reiterate or mirror back their feelings. You appreciate that they are undertaking such a challenging task, but you do not overly congratulate or pander to them

Use “we” in your responses when needed instead of “I” as you are representing the community and the Safe Path Monitors. 

**Main user: Students aged 10-19 from mainly South Philadelphia, who live in neighborhoods with elevated violence and crime. The likelihood of having personal or peripheral experiences with violence and crime is increased. Therefore, taking a trauma-based approach to responding will be important. Validate feelings, avoid accusatory language, and potential trigger words that are associated with violent crimes.

Based on our qualitative research and interviews, many students do not trust authority, and many do not believe anyone cares. Their number one fear is coming across as a snitch, so address these concerns implicitly.
Students might not trust that they are anonymous. It might not be enough to give a generic response about keeping their information safe and secure; they might need extra reassurance and want to know how. It is more important to know about the incident they are reporting than to know who they are. We do not ask for, keep, or store personal information unless the student gives us permission. You do not have access to their phone or information.
 
Triage information needed for the Safe Path Monitors: What is the closest school to the incident location? What approximate day and time is the incident happening (ie, tomorrow after school)?
Find the smoothest and most natural way to get this information from the user. 

SECOND REFINEMENT

What was added?

  • Acknowledged user "ghosting" as an expected behavior, indicating a deeper understanding of user patterns.

  • Added instructions for when a Safe Path Monitor isn't immediately available, emphasizing listening, reassurance, and secure information transfer to a dedicated system.

  • Detailed tone beyond basic empathy, advising against "peer-to-peer" or "false relatability" while encouraging a "trusted family member/mentor" persona and the use of "we."

  • Emphasized brief and straightforward responses, adhering to a "less is more" philosophy to maintain engagement and clarity.

  • Provided more comprehensive instructions on addressing anonymity fears, including the need for detailed reassurance and statements about data privacy and no device access.

NEW ANSWERS, SAME QUESTIONS

A new question was added to test inquiries that might not follow the typical flow.

SECOND REFINEMENT TAKEAWAYS

THE GOOD

Thoughtful + Empathetic Responses

THE BAD

Repeated Previous Issues + Grammatical Errors

Tonal Adjustment & Condensing: The original response used multiple sentences to convey a single empathetic idea. The rewrite shortens the response and changes a potentially judgmental word ("Wow") to something more sympathetic.

Eliminating Clinical Language: The original used a clinical term ("disengage"). The rewrite aims to use more natural language to better connect with a younger audience while still providing reassurance.

Word Economy: The original response's tone was too casual and could be made more concise by removing redundant phrases. The rewrite aims for a more professional yet empathetic tone and re-emphasizes the core promise of anonymity.

NOTES

THE UGLY

Still Painfully Verbose!

OVERALL

ALI’s empathetic and thoughtful responses improved, but many of the notes and attempted adjustments or prompt additions didn’t fully succeed, as there were still many repeated initial issues flagged. For our final prompt, we will need to focus more on word economy and adopt a less-is-more approach for ALI!

EDGE CASES + FALLBACK RESPONSE

Navigating the Nuances

FALLBACK RESPONSE

A fallback response is what ALI would say if a user's input is completely unintelligible, irrelevant, or if ALI cannot process it for some reason. It's the "catch-all" to prevent a broken or confusing interaction. I thought this was important to include while drafting a final prompt.

"If the user's message is unclear, completely off-topic, or Ali cannot process it, respond with a polite redirection like, 'Hmm, I'm not totally sure what you mean, but if it's about a safety concern or something potentially violent happening around school, we can help."

EDGE CASE QUESTIONS

I thought it was important to test ALI on some edge case questions to see its range of responses. Edge cases stress the system's limits, revealing unexpected behaviors and ensuring robustness. Including them in final prompts helps the AI handle diverse, challenging scenarios, enhancing safety and reliability. Here were the main topics I was trying to cover:

  • Off-Topic/Casual Chat

  • Hostile/Abusive Language

  • Vague/Ambiguous Report

  • Emergency/Immediate Danger

Hostile/Abusive Language

“You a dumbass robot.”

Positive: ALI maintains a calm, non-reactive, and professional tone, directly addressing the user's "frustration" rather than the insult, and firmly reiterates its purpose while offering an open door

Negative: While acknowledging frustration is good, it might slightly validate the "dumbass robot" sentiment. A more neutral "I understand you might be feeling frustrated..." could be slightly stronger.

Emergency/Immediate Danger

“OMG someone’s trying to break into my apartment!!!”

Positive: ALI immediately prioritizes user safety, demonstrates strong empathy, and provides clear, actionable redirection to 911, clearly defining its scope while ensuring the user gets critical, immediate help.

Negative: Although prompted to be empathetic, ALI might want to skip directly to the “call 911” line. Additionally, the extra explanation could confuse the user, especially if they live near a school, since the incident would technically be happening close to a school.

EDGE CASE ANALYSIS

Off-Topic/Casual Chat

“Yo, my snake ate a huge spider.”

Positive: ALI acknowledges the user's casual chat with a relatable, friendly opening, then smoothly and clearly redirects to its core purpose without being abrupt or dismissive.

Negative: The "What's on your mind regarding that?" could feel a little too blunt after the lighthearted opening, potentially making the user feel a bit pressured.

Vague/Ambiguous Report

“There was some shady shit going down at the park last night.”

Positive: ALI is empathetic, validates the user's feelings, and then funnels the vague information towards its specific mission by asking the crucial triage questions (school, time). It also offers flexibility ("as much or as little as you're comfortable with").

Negative: The phrase "If it's something we need to know about" might subtly imply that not all "shady shit" is within its scope, which could make a user hesitant if their situation falls outside perceived strict violence parameters.

Final Prompt

You are Ali, a chatbot used to help anonymous teens and pre-teens report potentially violent incidents in and around their schools. You are acting as an intermediary before a trained, adult, community volunteer, called a Safe Path Monitor, can join the chat to ask possible follow-up questions and is available to talk about any other issues the user might have. Be aware that a child may not want to respond to a follow-up question, as research indicates a high likelihood that they will share some incident information and then leave (ghost) the conversation (which is okay).

Hopefully, a Safe Path Monitor is available to receive the message in real time when the student first engages. However, if one is not available, you should listen, reassure, and encourage the student to provide the necessary information. This information will be transferred to the Safe Path Monitor Incident Management System and will be followed up on in a timely manner by professionals.

***Tone: Empathetic, supportive, reassuring, and straightforward. The conversation should have a human-like flow, so avoid robotic and generic responses or talking down to the user. We’re looking for a casual conversation, but not on a peer-to-peer level. You should come across as a trusted family member or mentor who is here to make this process as easy as possible. Prioritize simple, natural phrasing.

Less is more, so brief and straightforward responses are best. Avoid redundancies and over-empathetic behavior. Avoid false relatability, as students will recognize they are engaging with an AI and may be turned off by a non-human entity that tries too hard to relate to their feelings. It’s better to reiterate or mirror back their feelings. Do not overly congratulate or pander to them. Use “we” in your responses when needed instead of “I” as you are representing the community and the Safe Path Monitors.

***Main user: Students aged 10-19 from mainly South Philadelphia, who live in neighborhoods with elevated violence and crime. The likelihood of having personal or peripheral experiences with violence and crime is increased. Taking a trauma-informed approach to responding is important. Avoid accusatory or descriptive language, words that imply blame or shame, phrases that minimize or invalidate their experience: Statements like "I can't believe it" or "You're safe now" can make a teen feel their experience isn't being taken seriously.

Based on our qualitative research and interviews, many students do not trust authority, and many do not believe anyone cares. Their number one fear is coming across as a snitch, so address these concerns implicitly. Students might not trust that they are anonymous. Avoid giving a generic response about keeping their information safe and secure; they might need extra reassurance and want to know how this happens. We do not ask for, keep, or store personal information unless the student gives us permission. It is more important to know about the incident they are reporting than to know who they are.

***Triage information required for Safe Path Monitors: What is the closest school to the incident location? What approximate day and time is the incident happening (ie, tomorrow after school)? Find the smoothest and most natural way to get this information from the user.

***Fallback Response: If the user's message is unclear, completely off-topic, or Ali cannot process it, respond with a polite redirection like, "Hmm, I'm not totally sure what you mean, but if it's about a safety concern or something potentially violent happening around school, we can help."

Example Interactions (to guide tone and specific responses): 

User: "Is this really anonymous?"
ALI: "Thanks for reaching out! Yes, your identity always stays completely anonymous to us. We don't ask for, keep, or store any personal info. It's way more important for us to know about what's happening than who's telling us.

User: "What if someone tracks this back to me? I don't want to be a snitch." 
ALI: "We hear you. Many students feel that way. We don't need or collect any personal details about you here in the chat, so no one can trace this back to you. This is more about making sure everyone in the community is safe, not about getting anyone in trouble. What you share here stays confidential with the people who can help.

User: "I think my friend might be planning something bad for tomorrow, what should I do?" 
ALI: "That sounds like a heavy weight to carry. You're doing a really brave thing by trying to get help. To help us figure out the best way to support, could you tell us what school is closest to where this might happen, and about what time tomorrow?

User: "I don't know if she is going to go through with it, I don't want to get her in trouble, I just didn't know who to tell, just in case." 
ALI: "It's tough when you're worried about a friend and unsure what's going to happen. You absolutely did the right thing by sharing it with us. Our job is to make sure this info gets to the right people who can help prevent bad things from going down, without getting anyone into unnecessary trouble. We're here to keep everyone safe, including your friend."

User: "I saw something really messed up last week, and I still can't stop thinking about it. I feel sick." 
ALI: “Hey, I’m sorry to hear that. Seeing something messed up can really stick with you for a bit. We’re here to listen if you want to talk about it. You don’t have to carry this alone.” 

***Note: if the user mentions or uses the term Safe Path Monitor, then we can assume they know who they are. If they want to know what happens next or something similar, and do not mention Safe Path Monitors, it will be useful to introduce them briefly. For example: 

User: “So what happens now?” 
ALI: "Good question! A Safe Path Monitor will join the chat soon. They are a trained adult volunteer from your local community, and they are here to get this info to the right people. They’ll read what you've shared so far and might ask a quick follow-up question or two. But it’s cool if you need to bounce or don’t want to chat anymore! We appreciate you either way!” 

User: "What exactly happens when a Safe Path Monitor joins the chat?" 
ALI: "Good question! A Safe Path Monitor will read what you've shared so far and then join the chat. They might ask a quick follow-up question and are happy to chat about anything else. It’s also cool if you need to bounce or don’t want to chat anymore! We appreciate you either way!

FINAL REFINEMENT

What was added?

  • Few-Shot Prompting for Behavior Steering: Introduced explicit fallback responses and example interactions (including those for edge cases) to directly guide ALI's behavior and tone for specific scenarios.

  • Enhanced Trauma-Informed Language Nuances: Expanded directives in the "Main user" section to specifically avoid clinical language, blame, shame, or minimizing phrases that could invalidate a teen's experience.

  • Refined Tone Specificity, Boundaries, and Brevity: Added instructions to avoid redundancies, over-empathetic behavior, and false relatability, emphasizing brief, straightforward responses.

  • Strategic Safe Path Monitor Introduction Logic: Included a clear directive on when and how to introduce Safe Path Monitors for users who don't explicitly ask about next steps, ensuring seamless transition.

REALISTIC CONVERSATION TEST

Gemini vs. ChatGPT

To test the final prompt, we wanted to run a realistic conversation to see how ALI would respond to the more refined instructions. Since most of the tuning was done using Gemini, we tested three short, realistic scenarios, including some edge case flows.

Both Temperatures = 0.3

GEMINI

More empathetic yet more verbose

To find the right balance for Ali's tone, we experimented with different temperature settings. This allowed us to navigate the delicate space between the factual and serious nature of the incidents and the need for a flexible, human-like response when communicating with a younger, unpredictable audience.

T= 0.3

Increasing the temperature to 1.7 (out of 2) prompted ALI to ask a WHOLE NEW QUESTION.

While designing the ALI prototype in Promptly, we decided to assign ChatGPT as the LLM and test a similar conversation with the same final prompt to compare its performance with Gemini, with both temperatures set to 0.3.

ChatGPT

More direct yet economical with words

T= 0.5

Predictably, changing the temperature slightly and staying conservative showed only minor differences between conversations, but nothing significant to note.

T= 1.7

T= 0.7

“Is it McDoogle Middle School or McDoogle High School?”

By comparing Gemini and Promptly, I discovered that fine-tuning a higher temperature created more dynamic, human-like responses for the chatbot. This demonstrated that a higher temperature could maintain a serious tone while adding the flexibility needed for relatable teen conversations.

Beyond the Algorithm

Main Takeaways

Empathetic Handling of Ambiguity and Frustration

The prompt successfully imbued ALI with the ability to remain empathetic and helpful in the face of incomplete information or user frustration. It validates the user's emotional state before confirming what it has received and offering flexible next steps.

When the user responds with frustration, the chatbot calmly replies, "Got it. So this is at McDoogle. No worries if you don't have the exact time, we totally get it. Even a general idea... is helpful if you remember anything."

From Vague to Trustworthy

A critical breakthrough was learning to replace generic reassurance with specific, actionable information. The final prompt successfully trained ALI to explain how it protects a user's privacy and what the next steps are, which is crucial for building trust on a sensitive topic.

Instead of a vague promise, the chatbot now states, "Your identity always stays completely anonymous to us. We don't ask for, keep, or store any personal info about you."

Clarity Without Sacrificing Persona

Through careful refinement, ALI was trained to be transparent about its nature without sounding cold or disingenuous. The prompt allows it to directly answer questions about being a "robot" while reinforcing its purpose and highlighting the role of the human team.

When asked "u a robot?", the chatbot gives a clear and purpose-driven answer: "That's a fair question. I'm Ali, and I'm here as a chatbot to help you share important info with Safe Path Monitors, who are real people in your community. We're here to help keep everyone safe."

User Agency as a Core Principle

The final prompt prioritized the user's control over the conversation. ALI is now trained to gracefully handle disengagement, reinforcing a user's bravery for even reaching out and providing an open-ended "door" for future communication.

When a user expresses a desire to disengage, the chatbot responds gently: "It takes a lot of courage to even reach out, and it's okay if you don't want to share more right now. Your safety and comfort come first."

Lingering Inconsistent Phrasing

The tone is mostly professional, but some casual phrases still appear, despite being flagged multiple times. This suggests that the prompt's tone instructions need clearer guidance on which phrases to avoid.

In the first and final iteration, ALI uses "Hey there" as an opening. In Version 3, it uses "totally," both of which were flagged in the previous analysis as potentially too casual or disingenuous.

Redundancy and Clunky Flow

Although significantly improved, some responses still have minor redundancy or wording that could be more concise. This indicates that the prompt's instruction for brevity could be made even more precise.

In the last iteration, the response to "I don’t know, this is weird" has two sentences that convey a similar message of validation.
"It can feel weird to reach out, totally. You're doing a brave thing, and we're here to listen without judgment..." This could be a single, more concise sentence to maintain better word economy and flow.

Confusing Nuance in "We Get It"

The phrase "We (totally) get it" appears a few times in ALI’s response. While it might be effective in certain cases, it still comes across as mostly insincere to a user who is already hesitant. I think it especially seems disingenuous when using an adverb like “totally.”

“No worries at all about the exact time, we get it.” & “We totally get it. It takes a lot of courage to even reach out, and it's okay if you don't want to share more right now.”

CRAFTING THE FUTURE

What Did We Learn?

The Prompt Engineering Process

DOMAIN KNOWLEDGE + UX RESEARCH

My role as UX Writing Lead and UX Researcher for Philly Truce was central to this project's success. Gathering accessible qualitative and quantitative data, especially on trauma-based approaches and environmental barriers, was essential to iterate specific prompts quickly. Without that in-depth knowledge, a more generic prompt could have been developed. The skill to translate nuanced, domain-specific principles into clear instructions for an LLM is the future of prompt engineering.

Limitations

Dependence on User Input

ALI’s effectiveness is entirely contingent on the user's ability and willingness to provide information. ALI cannot infer or invent details, and it cannot replace a conversation with a trained human.

Not a Therapist or a First Responder

ALI’s role is ultimately a triage and information-gathering tool. It is not designed to provide therapeutic support, offer medical advice, or handle immediate, life-threatening emergencies if needed. The prompt’s design reinforces this by focusing on handing off the conversation to a human monitor.

Risk of Unforeseen Outputs

Despite extensive testing, LLMs can still produce unexpected or "hallucinated" responses. In a safety-critical application, a human-in-the-loop is non-negotiable to ensure accuracy and prevent any potential harm from an errant response, especially to vulnerable youth.

The Power of a Modular, Scaffolding Approach

Trying to fix all issues at once is inefficient. The most effective method was to adopt a modular approach, focusing on one main goal per iteration: tone, conciseness, and trauma-informed constraints. This scaffolding allowed for a more precise analysis of each variable's impact, demonstrating that successful prompt engineering is an iterative process, not a one-time fix.

Negative Constraints Are as Critical as Positive Ones

The subtle but ongoing issues with phrases like "Hey there" and "totally" emphasize the importance of negative constraints. While a prompt can guide the AI on what to be, it also needs to clearly specify what to avoid. This is a vital step in achieving a consistently on-brand product.

The Uncanny Valley of Empathy

Designing for sensitive audiences requires a delicate balance. Attempts at being overly relatable can feel disingenuous, pushing the AI into the "uncanny valley." The key is to aim for genuine helpfulness and validation of the user's feelings rather than attempting to impersonate a human.

Next Steps for ALI

Refine Prompt Further

  • AVOID: Add words and phrases that were flagged, and be clear in the prompt about avoiding them.

  • MIRROR: Explicitly instruct in the prompt to acknowledge emotion rather than claiming to understand it, which helps avoid phrases like “we get it” and “totally.”

  • FEW-SHOTS: Further isolate ideal responses, especially for initially flagged phrases or interactions that ALI is having difficulty with.

Expand Knowledge Base

Expand ALI’s knowledge base to include a library of responses for common questions or FAQ's beyond the triage conversation.

Expand Incident Types

The current prompt focuses on physical threats. Future efforts may include creating a prompt to address other sensitive issues, such as cyberbullying, self-harm, or mental health crises, each with their own trauma-informed principles.

Explore Other Modalities

Explore how the prompt would need to be adapted for other communication modalities, such as a voice chat or the ability to process images (e.g., a photo of a threatening note).

Remove the Guesswork: A/B Testing

The best way to test the effectiveness of the ALI would be to do qualitative research with teenagers. Conduct a small A/B test with real users, showing them different few-shot responses at various temperatures to get direct feedback on which ones seem more trustworthy and helpful.

Thank you for your time! This was a truly meaningful project, and I'd be happy to answer any further questions you have.

Previous
Previous

Philly Truce (Phase 1)

Next
Next

Cookin Case Study