Human–Computer Interaction • Machine Learning • Technology Policy

I design human-centered ML systems — decision support people can understand, challenge, and trust.

I’m Yogesh Luitel — a computer science instructor and early-career researcher. Technology is never neutral. We research what it does to people.

Current
CS Instructor (Grades 11–12)
Xavier International College • Kathmandu
Research
Human–AI interaction
Trust calibration • uncertainty • evaluation
Policy lens
Accountable deployment
Privacy • auditability • governance

About

I work at the boundary between research and practice: human–AI interaction, empirical ML evaluation, and technology policy for accountable deployment.

Short bio

I’m interested in how people make decisions with ML systems: what they see, what they trust, and what they can contest. I focus on human-centered design patterns (uncertainty cues, explanations, feedback loops) paired with evaluation methods that go beyond accuracy.

My research hook

Turning predictions into decision support people can calibrate: evidence + uncertainty + accountability — not just scores.

Research interests

Human–AI interaction Trust calibration Uncertainty-aware UX Explainability Empirical evaluation Robustness & failure modes Transparency & governance Privacy-by-design
  • HCI: interface patterns for sensemaking, feedback, and contestability.
  • ML: evaluation, subgroup errors, robustness, interpretability.
  • Policy: accountability constraints for real institutional deployments.

Selected projects

Polished, research-aligned projects: interface design + evaluation + real-world constraints.

HCI • XAI

Human-Centered ML for Programming Misconceptions

Decision Support

Teacher-facing decision support that surfaces likely misconceptions with uncertainty cues and explanations — designed to reduce over-trust and preserve instructional agency.

  • Interface: uncertainty signaling + evidence views + easy override
  • Evaluation: behavior metrics (time-to-decision, correction patterns)
  • Goal: support sensemaking, not automated grading
Python Human-in-the-loop Explainability Evaluation
Robustness • Reproducibility

ML Evaluation & Interpretability Toolkit

Audit-ready

A reproducible evaluation pipeline that turns “good accuracy” into deployable evidence: subgroup errors, robustness checks, and interpretability reporting.

  • Evaluation: metrics + subgroup slices + failure modes
  • Interpretability: model-agnostic explanation views
  • Use case: institutional review & governance
scikit-learn SHAP Error analysis Reporting
Empirical ML • NLP

GenAI Usage & Learning Outcomes Study

Methods-first

Empirical analysis of how generative AI usage patterns relate to learning outcomes using structured data + text features, emphasizing careful claims and reproducible methodology.

  • Data: usage logs + text signals
  • Methods: interpretable modeling + effect-size-grounded validation
  • Outcome: manuscript in preparation (share on request)
TF–IDF Feature engineering Stat tests Reproducibility
If a project is private, I can share a short demo clip or 1-page writeup by email.

Research

Where I want to go next: HCI + ML evaluation + policy-aware deployment.

Current focus

  • Human–AI interaction: decision-support UX, trust calibration, contestability.
  • Human-centered ML: uncertainty cues, explanation interfaces, feedback loops.
  • Empirical evaluation: subgroup errors, robustness, failure modes, reproducibility.
  • Technology policy: transparency, privacy, governance constraints.

Drafts & directions

  • GenAI usage → outcomes: empirical ML + NLP methodology (in preparation).
  • Accountability for ML-assisted decisions: uncertainty + oversight design patterns.
  • Low-resource policy constraints: privacy-by-design and evaluation standards.
Reach out about collaboration

Experience

Teaching + research + building — grounded in real users and constraints.

2024–Present
Computer Science Instructor • Xavier International College

Teach Python, DBMS, and software development; lead labs and supervise projects; design digital learning materials and workflows.

Python DBMS Mentoring Learning workflows
2023–2024
Undergraduate Research Assistant • Phoenix College of Management

Studied adaptive learning systems and built ML-based student performance analysis to identify conceptual weaknesses; supported research design and evaluation.

Learning analytics Mixed methods ML evaluation Adaptive learning

Contact

If you want to talk research, collaboration, or opportunities — email is best.

Links

I’m especially interested in opportunities where interface design, evaluation rigor, and governance matter as much as the model.