Adversarial Prompt Expert
Up to $80/hr
Depending on the project
Bachelor's
Remote
Part-time
Flexible hours
Overview
You’ll be part of a red teaming project focused on probing large language models for failure modes and harmful outputs. Your work will involve crafting prompts and scenarios to test model guardrails, exploring creative ways to bypass restrictions, and systematically documenting outcomes. You’ll think like an adversary to uncover weaknesses, while collaborating with engineers and safety researchers to share findings and improve system defenses.
Details
The position is remote and asynchronous - work independently from wherever you are.
Flexible hours and the ability to work remotely, with a commitment of approximately 10 to 20 hours per week.
Project work includes and is not limited to developing domain-specific prompts and evaluating LLM responses.
Dedicate time researching topics that are interesting to you, with the assistance of AI.
Learn new skills while contributing to the changing world of AI across various disciplines.
Placement into a project will be dependent on project availability.
Qualifications
Heavy LLM Usage — hands-on experience with multiple models (open- and closed-source), comfort experimenting across systems.
Prompt Engineering & Jailbreaking — skill in crafting prompts, evasion techniques, and creative ways to bypass restrictions.
Adversarial / Security Mindset — ability to think like an attacker, with bonus points for any red teaming or offensive security background.
Persistence & Creativity — willingness to try many variations, think outside the box, and push edge cases.
Clear Documentation — ability to log attempts and outcomes systematically, and communicate issues clearly.
Ethical Awareness — understands boundaries and handles sensitive content responsibly.This program is open to U.S.-based students, candidates, and recent graduates with US work authorization.
The application process