Key Takeaways from the MedBiquitous Emerging Technologies Conference
Artificial Intelligence (AI) in medical education has spent the last few years in a familiar place: strong prototypes, promising pilots, and plenty of debate about where (and whether) it belongs. At the Association of American Medical Colleges’ (AAMC) MedBiquitous Program's 2026 Emerging Technologies for Teaching and Learning – Digital Demonstrations Virtual Conference, the conversation felt noticeably more grounded. Across two afternoons of live demonstrations, the emphasis shifted from "What can AI do?" to "How do we operationalize AI inside real educational workflows—safely, consistently, and defensibly?"
That subtext also showed up in the depth of implementation experience shared. More than a dozen sessions were presented by Elentra member institutions, reflecting a community that isn’t simply experimenting with new tools, but actively shaping how emerging capabilities are integrated into curriculum, assessment, evaluation, and reporting.
There were six themes that appeared repeatedly across the many sessions of the conference, suggesting a common set of goals among programs leveraging emerging technologies:
-
Turning feedback into consistent, usable guidance
AI is being used to improve the consistency, clarity, and usefulness of assessment and narrative feedback—so it can reliably support coaching and decisions, not just get stored and forgotten. -
Making large volumes of comments easier to act on
AI helps education teams quickly sort, summarize, and quantify large volumes of qualitative comments so they can focus on the themes that matter most and act faster. -
Using AI to strengthen reasoning—not replace it
The most responsible tools are designed to prompt reflection and strengthen clinical reasoning, rather than simply providing answers that encourage shortcut thinking. -
Drafting high-stakes documentation safely and transparently
AI can speed up drafting of summative narratives, but the safest models embed structured review steps and clear links back to the evidence behind the final wording. -
Scaling faculty development with practical, personalized support
AI-enabled resources make it easier to support busy, distributed educators through short, practical learning moments, low-stakes practice, and personalized development pathways. -
Moving from AI pilots to sustainable, governed programs
Institutions are moving beyond one-off experiments by putting governance, scope, and ongoing evaluation in place so AI tools can be trusted and sustained.
Below, we expand each theme with what was demonstrated, why it matters, and what it signals for health professions education teams planning their next steps.
Turning feedback into consistent, usable guidance (not just more narrative text)
Narrative evaluations remain one of the most important—and most variable—signals in health professions education. They capture nuance, context, and lived observation that rating scales can miss. But they also introduce recurring challenges: inconsistent quality, vague language, bias risk, and high review burden.
One of the clearest "maturity signals" across the conference was how many solutions were not focused on generating more text. Instead, they treated feedback as something that requires a quality system—a repeatable, auditable process that improves clarity and actionability over time.
Common design patterns showed up across multiple demonstrations:
- Rubric-driven evaluation of narrative quality. Several tools formalized what "good narrative feedback" looks like (specificity, behavioral evidence, alignment to competencies/objectives, usefulness for coaching) and used that rubric to score, flag, or improve comments.
- Evidence visibility. Rather than producing an untraceable summary, stronger approaches highlighted the phrases and observations that supported a theme or conclusion. This matters for educator trust and for defensibility when narrative informs progression decisions.
- Human-in-the-loop calibration. The most credible demos positioned AI as a first-pass reviewer—triaging, suggesting edits, and flagging risk—while educators remained accountable for final interpretation and decisions.
Another cluster of sessions applied similar thinking to structured clinical documentation and performance tasks: rubric-aligned grading, feedback drafting, and consistency checks designed to reduce variance while preserving educator control.
Why it matters: high-quality feedback is one of the most cost-effective interventions in education, but it only works when it is specific, credible, and actionable. The conference showed AI being used less as a writing tool and more as a feedback quality amplifier—one that helps programs scale consistency without losing professional judgment.
Making large volumes of comments easier to act on (from "too much to read" to clear priorities)
If there was one "quiet hero" use case across the conference, it was the idea of helping educators and administrators manage the sheer volume of qualitative data generated by modern programs.
Course evaluations, rotation feedback, narrative assessments, reflective writing, committee reviews—many health professions education disciplines produces an extraordinary amount of text. The bottleneck is rarely a lack of information. The bottleneck is time: time to read, code, synthesize, compare across cycles, and translate insight into action.
Several demonstrations showed "analysis agents" built to absorb large datasets of comments and return outputs designed for committee workflows—not just summaries. The strongest versions went beyond "here’s what people said" and instead provided:
- Theme clustering with clear labels
- Frequency/counts (what’s most common, what’s emerging, what’s fading)
- Comparisons across time or cohorts ("what changed since last cycle?")
- Outputs shaped for decision-making (prioritization, risk signals, and next-step framing)
Why it matters: continuous improvement only becomes a cultural norm if it becomes operationally feasible. AI is increasingly being applied as an "overload relief layer" that helps programs listen at scale—and act faster.
Using AI to strengthen reasoning—not replace it (designing for judgment over shortcuts)
As generative tools become ubiquitous, a central educational question is no longer hypothetical: What happens to clinical reasoning when instant responses are always available?
A particularly thoughtful thread across the conference focused on tools intentionally designed not to act as answer engines. Instead, these systems functioned as coaching supports—prompting reflection, surfacing thinking patterns, and guiding learners through structured reasoning frameworks.
The practical distinction is important:
-
Answer-oriented tools risk accelerating shortcut behavior and cognitive offloading.
-
Judgment-oriented tools use AI to make reasoning visible, interrogate assumptions, and build metacognitive habits.
Across several demos, a recurring strategy was to structure interactions around reflective prompts—asking learners to articulate rationale, consider alternatives, identify uncertainty, and recognize bias patterns. Some implementations also emphasized constrained knowledge boundaries (e.g., limiting outputs to curated materials) and traceable citations to reinforce trust and reduce hallucination risk.
Why it matters: "using AI" is not a learning strategy. The instructional design around AI determines what learners practice—and what habits are reinforced. The conference signaled a shift toward designing AI experiences that protect and develop judgment, rather than merely improving access to information.
Drafting high-stakes documentation safely and transparently (review gates, traceability, and accountability)
Another consistent theme was the modernization of high-stakes learner documentation and summative narratives. These outputs require programs to synthesize large evidence bases into coherent, defensible language—often under intense time pressure and with meaningful equity implications.
The most responsible demonstrations converged on a shared pattern:
-
Aggregate evidence across assessments and narrative evaluations
-
Extract salient themes and "notable characteristics" in a structured way
-
Draft narrative language that follows consistent headings and tone expectations
-
Route through defined review gates (advisor, committee, or program-level review)
-
Preserve traceability from final language back to supporting evidence
What stood out is that the "review gate" wasn’t treated as an optional step; it was framed as the core safety feature that makes assistive drafting viable in high-stakes contexts. Tools that generate text without governance may save time, but they also increase risk. Tools designed with workflow, oversight, and auditability can improve efficiency while maintaining accountability.
Why it matters: high-stakes documentation is exactly where programs need both speed and defensibility. The conference reflected a growing understanding that the safest path is not "AI writes the document," but "AI supports a structured drafting-and-review process with transparent linkage to evidence."
Scaling faculty development with practical, personalized support (microlearning, rehearsal, and tailored pathways)
Faculty development has always faced structural constraints: limited time, distributed teaching environments, and finite support staff. Several sessions demonstrated how AI is being used to reshape faculty development into something more continuous and operational.
Three patterns appeared frequently:
-
Microlearning that fits real schedules
Instead of long workshops that compete with clinical and teaching responsibilities, microlearning resources were designed to be consumed quickly, revisited easily, and embedded into real workflows. This includes multi-modal formats (visual, audio, short video), accessibility-conscious design, and feedback loops to improve content relevance. -
Rehearsal environments for difficult teaching moments
Some of the most compelling applications used simulation-style AI interactions to let faculty practice high-stakes conversations—especially around feedback—without the risk of "getting it wrong" in front of learners. The best designs paired practice with a rubric and structured reflection. -
Personalized development pathways
Rather than generic modules, tools generated individualized development plans grounded in structured input data, self-reflection, and curated resources—moving faculty development from "one-size-fits-all programming" toward guided, actionable roadmaps.Why it matters: scaling faculty development isn’t just about content volume; it’s about repeatability and relevance. AI is increasingly being applied as a way to deliver consistent, practical, personalized support—especially for institutions with distributed educators and limited centralized capacity.
Moving from AI pilots to sustainable, governed programs (scope, guardrails, and continuous evaluation)
A visible slice of the conference focused on custom-built assistants tailored to educator and administrative workflows: aligning objectives, drafting assessment items, supporting teaching feedback, and accelerating content creation.
What made the strongest sessions stand out wasn’t the sophistication of the prompts—it was the seriousness of the operating model behind them. Presenters emphasized that these tools only become institutional assets when there is discipline around:
-
Scope definition: what the assistant will and will not do
-
Data handling: privacy, de-identification, and boundary-setting
-
Output constraints: templates, rubrics, and review requirements
-
Monitoring and evaluation: quality checks, drift awareness, and continuous improvement
-
Documentation: traceability, decision accountability, and auditability
In other words, the tools that seemed most "ready for scale" treated governance as a product feature, not an afterthought.
Why it matters: health professions education disciplines typically represent a high-stakes environment. If AI is going to live inside core workflows, programs need more than clever prototypes—they need repeatable, supportable, governable systems.
Summary
Across the conference, a unifying “workflow equation” kept emerging:
Unstructured inputs at scale → structured frameworks → AI-assisted synthesis → human review → action + reporting
That pattern showed up in assessment feedback, course and clerkship improvement, faculty development, simulation coaching, and high-stakes documentation. The conference’s strongest signal was not that AI is “coming” to health professions education; it was that AI is already being operationalized. For academic program teams planning their next steps, this conference reinforced the idea of starting with a high-friction workflow, adding structure, designing for review, and measuring impact in operational terms. That’s how AI will move from “interesting” to sustainable.
At Elentra, we learn from and support the institutions doing this work thoughtfully. If you’d like to explore how Elentra helps connect curriculum, assessment, evaluation, workflows, and reporting in one platform—contact us today.