Applied

I translate research on trust and human judgment into frameworks that can inform product decisions, measurement systems, and AI evaluation. I’m especially interested in how platforms and AI systems can incorporate human input without over-relying on noisy or unstable signals.

AI evaluation & human judgment

Dimensions of quality
Building and validating evaluation dimensions that users can judge reliably in context, and using disagreement as information rather than noise.
Examples: response quality dimensions, judgment stability, trade-offs across dimensions.
When to trust human feedback
Identifying when human judgments are reliable enough to steer system behavior, versus when they should remain advisory or be down-weighted.
Examples: signal strength, missingness patterns, calibration checks, robustness.

Trust & platform design

Reputation systems and marketplace trust
Applying research on reputation, social similarity, and selection to inform how platforms design trust signals and reduce unequal outcomes.
Examples: signal design, boundary conditions, unintended consequences.
Trust dynamics in AI-mediated interaction
Translating a dynamic view of trust thresholds, control, and safety infrastructure into concrete system behaviors (e.g., when the system should be cautious, when it can be more autonomous).
This connects to my ongoing model work described on the Research page.

How I like to work

I work best when research questions are tied to a clear decision context (what will change if we learn X?). I prefer simple, testable frameworks; careful measurement; and transparent assumptions. I’m comfortable moving between theory and implementation, and between academic and product audiences.

Collaboration

If you want to discuss collaborations, talks, or interviews, see Contact.