AINeutralarXiv โ CS AI ยท 5h ago6/10
๐ง
Bring Your Own Prompts: Use-Case-Specific Bias and Fairness Evaluation for LLMs
Researchers present a decision framework and open-source library (langfair) for evaluating bias and fairness risks in Large Language Models across specific deployment contexts. The study demonstrates that fairness evaluation cannot rely on benchmark performance alone, as risks vary substantially depending on use case, prompt characteristics, and stakeholder priorities.