Data Analyst | Product Analytics, Pricing Optimization, and Machine Learning
I analyze product and operational data to drive pricing decisions, improve conversion, and optimize efficiency. I have built optimization models for shipping costs, end-to-end ML pipelines, and analytics frameworks that translate directly into business impact.
My focus is on solving high-leverage problems where data informs strategy and drives measurable outcomes.
I’m a graduate student at Purdue University (MS in Business Analytics, Aug 2026) who cares about one thing: what should we do next?
At Jambo Club I designed geo-split experiments that cut CPC by 27% and built survival models that doubled retention. At Mayacrew I ran cohort analyses that lifted conversion 15% and automated B2B dashboards that saved 40% of reporting time. At Purdue Data Mine I built KPI dashboards and forecasting models for corporate partners.
I work across the full analytics lifecycle — structuring messy questions, pulling and cleaning data, building models, and delivering recommendations that stakeholders act on.
Built an interactive Plotly dashboard centralizing corporate KPIs, backlog metrics, and a reorder likelihood signal flagging high-risk and high-opportunity products. Identified repeatable inventory and demand patterns at the SKU level using time-series analysis and clustering, improving sales predictability and guiding replenishment strategy.
Owned marketing performance and retention analytics across paid acquisition and in-product funnels. Designed geo-split experiments and multivariate bid tests, applying regression-based lift analysis to reduce CPC by 27% while maintaining conversion volume. Built retention curves and survival analyses that identified critical churn moments, guiding features that increased retention from 22% to 53%.
Assessed acquisition effectiveness and segment economics through funnel and cohort analyses, delivering strategic recommendations that increased digital conversion rates by 15%. Built Tableau and Excel dashboards for B2B engagement KPIs, reducing manual reporting time by 40% and shifting budget toward the highest-ROI B2B segments.
Identified seasonal and macro-driven demand patterns in semiconductor equipment utilization, linking utilization shifts to customer CapEx behavior. Developed an ensemble forecasting model combining Random Forest and XGBoost, improving utilization prediction accuracy by 13%.
Contributed to more than 500% growth in student enrollment over two years through academic consultations and trial classes. Led and mentored a team of instructors, coordinating lesson planning, sharing best practices, and monitoring instructional quality.
Business Analytics & Information Management
Purdue University, Daniels School of Business
Relevant Coursework: Business Analytics, Financial Analytics, Spreadsheet Optimization, Data Mining
Business Analytics & Information Management
Purdue University, Daniels School of Business
Dean’s List: Fall 2022, Spring 2023, Fall 2023, Spring 2024 · Graduated with Distinction
Each project follows Problem → Approach → Results → Business Insight
Problem: A baby-gear e-commerce company shipped 17 stroller models in 17 different box sizes. Carriers charge by dimensional weight — (L × W × H) / 250 — whenever it exceeds actual weight. Oversized boxes were inflating DIM charges on nearly every shipment, and the cost was either absorbed as margin loss or passed to customers through higher shipping fees, hurting checkout conversion. At the same time, stocking 17 box SKUs created warehouse picking complexity and storage overhead.
The core tradeoff is non-linear: a 6.5% cost increase buys a 59% reduction in box SKUs. That ratio gives operations a clear, quantified decision point rather than an open-ended debate about simplification. The two outlier products are responsible for an outsized share of DIM cost and should be evaluated separately — whether through redesigned packaging, a flat shipping surcharge, or a free-shipping threshold set above their DIM break-even point. This analysis turns a vague “shipping is too expensive” complaint into specific, actionable levers: which boxes to stock, which products to reprice, and exactly how much each decision costs.
Problem: Sales and operations teams relied on gut feel to decide which products to restock. There was no shared view of backlog health or early demand signals, leading to frequent stockouts on high-value SKUs.
Demand signals already lived in historical order data — the gap was visibility and alignment, not collection. A scoped dashboard turned tribal knowledge into a repeatable planning rhythm and reduced reactive firefighting on high-value SKUs.
Problem: The company was scaling ad spend without knowing which channels actually drove incremental sign-ups. At the same time, only 22% of new users were still active after 30 days.
Churn concentrated in two narrow windows, not a slow leak — so timely, targeted nudges beat broad blast campaigns. Rigorous geo tests gave finance confidence to reallocate budget toward channels with provable incremental lift.
Problem: Sports media and bracket analysts rely on subjective rankings to predict which 68 teams make the NCAA tournament and how they are seeded. Can a data-driven pipeline match or beat that intuition using only regular-season performance?
Schedule quality and late-season momentum mattered more than raw win totals — mirroring how selection committees weight “who you beat” and how you finish. Treating selection vs. seeding as separate decisions avoided a single model conflating two different objectives and improved both outputs.
Problem: Applied Materials planned equipment capacity using trailing averages. When customer demand shifted — driven by seasonal cycles and CapEx timing — utilization forecasts lagged behind, leading to over- or under-provisioning.
Customer CapEx timing was a stronger leading indicator of utilization than the equipment’s own trailing trend alone. Feeding that signal into forecasts shifted planning from reactive backfill to forward-looking allocation.
Problem: Lenders and investors need to flag companies at risk of bankruptcy before it happens. The challenge: financial data is messy — 64 ratios with over 40% missing values — and bankrupt firms represent less than 5% of observations.
A score beats a blunt label: risk teams can set thresholds to match portfolio policy and expected loss appetite. Thoughtful imputation and feature integrity moved the needle more than chasing marginal gains from model complexity alone.
I focus on understanding the real-world problem before jumping into data. Domain knowledge is critical for asking the right questions and avoiding misleading conclusions.
I define clear success metrics that align with business goals, whether it’s revenue, conversion, cost, or efficiency.
I apply statistical methods and machine learning to analyze patterns, but I treat models as tools to support decisions, not replace them.
In an era where AI can generate instant outputs, the ability to evaluate results critically and apply domain knowledge is increasingly important. I focus on interpreting results in a way that makes sense for the business.
My goal is not just analysis, but clear recommendations that stakeholders can act on to improve outcomes.
I'm always open to new opportunities and collaborations