I built and owned the data pipeline and automated report-generation workflow for Columbia Business School’s 360° Leadership Feedback program, enabling personalized, multi-source PDF reports at scale for MBA and EMBA cohorts.

The workflow replaced a costly third-party reporting setup, improved interpretability for students and coaches, and was recognized with a school-wide teaching award.

Context & Problem

As part of Columbia Business School’s core leadership curriculum, MBA and EMBA students complete a 360° feedback assessment that combines self-ratings, peer evaluations from classmates, and feedback from current or former coworkers. The resulting report serves as a key input to executive coaching sessions, where students reflect on their leadership style and define development goals.

Prior to Fall 2022, the reporting workflow relied on a third-party Qualtrics XM solution that generated basic summary statistics and offered limited flexibility in how feedback could be structured or interpreted. While data collection was robust, the reports themselves were often difficult to use in practice. Comparisons were coarse, insights were hard to interpret, and faculty and coaches had limited ability to tailor outputs to pedagogical goals.

At the same time, the program operated under strict constraints: fixed academic timelines, large cohorts, and the need to deliver sensitive, multi-source feedback in a way that was clear, psychologically safe, and usable by a non-technical audience. The challenge was not collecting more data, but transforming existing data into interpretable, decision-ready feedback that could reliably support leadership coaching at scale.

My Role & Ownership

I owned the data pipeline and automated report-generation workflow for the 360° Leadership Feedback system, from initial design through production. I built the reporting codebase from scratch, designed the data processing and automation infrastructure, and made core decisions about how multi-source feedback was processed, summarized, and presented to students.

Beyond implementation, I partnered closely with Management Division faculty and the Bernstein Center for Leadership to translate pedagogical goals into interpretable, actionable feedback. This included decisions about which comparisons to surface, how to balance clarity with analytical depth, and how to present sensitive peer feedback in a constructive way.

Over time, I evolved the system into a largely hands-off, fully automated workflow that reliably generated and distributed personalized PDF reports to thousands of students on a fixed academic timeline. I documented the system and structured the code to enable a smooth handoff to a dedicated data science team when I transitioned off the project.

Impact Snapshot

The Product: Personalized 360 Reports

How the reporting system translated multi-source data into usable leadership insight.

The reporting system was designed to translate complex, multi-source feedback into insight that students and coaches could actually use. The structure intentionally surfaces high-level patterns first, with the option to drill down into more detailed views when needed, balancing clarity, psychological safety, and analytical depth.

The Lead 360 Survey

Designed to collect aligned, multi-source feedback that supports meaningful self–other comparison while preserving rater anonymity and trust.

  • Self-assessment: During the week-long leadership course, students complete a self-assessment, rating their own leadership behaviors across a shared set of attributes.
  • Coworker feedback: Students nominate current or former coworkers, who complete an individualized version of the same survey, rating the student on the exact items used in the self-assessment.
  • Classmate feedback: Classmates, primarily within learning teams, provide peer evaluations informed by intensive teamwork, projects, and leadership exercises.
  • Privacy safeguards: Reports are generated only when a minimum of three raters per category are collected, preserving anonymity and encouraging candid feedback.
Example Lead 360 survey item in Qualtrics
Example of peer-report survey on students' Motivation and Vision.

High-Level Feedback Overviews

Surfaces clear, interpretable patterns to help students quickly calibrate how they see themselves versus how others experience them.

  • Self vs. others’ perceptions: One overview compares how students rated themselves with how they were rated by their evaluators. This view helps students assess how accurately they understand the impact of their behavior on others—an especially important insight for leaders working to calibrate their presence.
  • Calibration and reassurance: For many students, this comparison reveals that they are viewed more positively by others than they view themselves. Identifying the specific domains where this is most pronounced helps them self-reflect as they move forward.
  • Peer benchmarking for development: Another overview compares a student’s average evaluator ratings with the class average across leadership domains, highlighting relative strengths and areas for growth.
  • Actionable next steps: Students frequently use this benchmarking to guide development decisions, such as selecting electives or seeking leadership experiences that address lower-scoring domains (e.g., negotiation and cooperation).
Overview of self vs. others' perceptions
Example of a student's overview of their self vs. others' perceptions.

Domain-Level Deep Dives

Enables focused exploration of specific leadership domains once high-level strengths and gaps are identified.

  • Domain-specific views: Students can explore each leadership domain independently, with ratings broken out by source (self, classmates, coworkers) and compared against peer benchmarks.
  • Source-level transparency: This breakdown allows students to see where perceptions converge or diverge across different audiences, supporting more precise interpretation than a single aggregate score.
  • Item-level detail on demand: Within each domain, students can drill down to individual survey items, viewing item-by-item ratings by rater group.
  • Targeted development planning: After identifying priority domains, students use item-level feedback to pinpoint specific behaviors to work on (e.g., addressing a particular aspect of perspective-taking rather than the domain broadly).
Example of a domain-level deep-dive into Perspective Taking
Example of a domain-level deep-dive into Perspective Taking.

General Impressions & Personality (Big Five)

Provides broader context for leadership feedback by highlighting how personality traits are perceived across different audiences.

  • Broad impression benchmark: In addition to leadership-specific behaviors, the report includes feedback on the Big Five personality dimensions. These impressions provide a broader reference point for how students see themselves—and how they are seen by others—on traits that are not inherently “good” or “bad” for leadership.
  • Awareness over optimization: The goal of this section is not to score highly on any single dimension, but to increase awareness of how one’s personality is perceived across different audiences. Differences between self-ratings and others’ impressions often surface useful blind spots or strengths that shape leadership style indirectly.
  • Convergence and divergence: Students can see where their self-perceptions align with others’ impressions, and where they diverge, helping them interpret leadership feedback in a broader psychological context.
Example of a personality impressions
Example personality impression view comparing self-ratings with others’ impressions.

System Architecture & Automation

I designed a scalable analytics and reporting pipeline that transformed raw, multi-source survey data into reliable, personalized feedback products delivered on a fixed academic timeline. The system was built to minimize manual intervention, enforce privacy constraints, and produce consistent outputs at scale.

The reporting system was built as a two-stage analytics pipeline: a centralized data processing script followed by a parameterized report-generation layer.

Responsible AI for Qualitative Feedback

Each reporting cycle included tens of thousands of open-ended peer comments. These comments needed to be reviewed for offensive, inappropriate, or harmful language before being shared with students, both to protect recipients and to meet institutional standards for psychological safety.

By constraining AI to a clearly defined, high-friction task and embedding it within a human review workflow, the system balanced efficiency gains with accountability and trust.

Impact & Reach

Continuity & Handoff

Reflections & Tradeoffs

Recognition & Stakeholder Trust