Analytical Engineer who enjoys tackling challenging
data-related problems using Python to create applications
and machine learning (ML) algorithms to help turn
unrefined data into useful business decisions.
about me:
I am an aspiring Data Analyst with strong skills in Python, SQL, and Power BI, with a focus on data modeling, predictive analytics, and process automation. Through academic projects and independent analytical work, I have developed experience in transforming complex operational and business data into meaningful insights that support data-driven decision-making. My work includes data cleaning, exploratory data analysis, and building interactive dashboards that help visualize trends and performance metrics. I also have a strong interest in statistics and enjoy applying statistical methods to better understand patterns, relationships, and variability within data.
I am particularly interested in building structured and reliable data solutions while maintaining high standards of data integrity, accuracy, and analytical transparency. With a strong analytical mindset and attention to detail, I aim to bridge the gap between technical data analysis and practical business applications. I am continuously expanding my knowledge of modern data tools, statistical techniques, and analytical methodologies in order to solve complex problems and create measurable value through data-driven insights.
As a Master’s candidate with a foundational background in Industrial Engineering, I am dedicated to bridging the gap between technical operations and financial strategy. Having completed two semesters of advanced graduate study, I have developed a robust understanding of the financial landscape through core modules including Asset Management, Commercial Banking Management, and Marketing.
I have developed a particular specialization in Quantitative Methods in Finance. This area represents the intersection of my passion for statistical modeling and complex financial decision-making. My goal is to leverage my analytical rigor and engineering precision to provide high-level insights in quantitative analysis and financial management.
During my undergraduate studies, I built a comprehensive foundation in statistical analysis and probability, focusing on how these mathematical principles can be applied to optimize real-world industrial systems. I graduated with a 2.1 mark, balancing a rigorous curriculum with active leadership and research.
Academic Leadership & Instruction
Beyond my coursework, I was selected to mentor my peers and junior students in highly technical subjects:
For my thesis, I utilized Machine Learning and Data Mining methodologies to model and predict the Bullwhip (whiplash) effect across industrial supply chains, aiming to reduce inefficiencies through predictive analytics.
The automotive industry demands consistent quality under tight deadlines. I focused on bridging the gap between theoretical standards and daily production, ensuring our processes were both compliant and practical.
Managing the high-integrity data required for three successful IATF audits is what initially drew me to quantitative analysis. I learned how to identify patterns and risks within complex industrial systems—a skill I am now advancing through Quantitative Methods in Finance in my Master's program.
In the high-pressure environment of automotive manufacturing, I recognized that our manual risk-tracking system was a significant liability. The process was not only slow but carried a high probability of human error due to the sheer volume and complexity of the data involved. In response to the risks of a manual tracking system in automotive manufacturing, I designed an application to automate the entire process. This solved the issues of speed and human error that were inherent in managing such complex and high-volume data.
The Problem: A Fragile Manual Workflow
Before this automation, our quality tracking relied on a fragmented, manual chain that was extremely
difficult to maintain accurately:
I identified a critical liability in our FMEA tracking: a complex, manual system of data silos that relied on transcribing hundreds of data points between Excel and Word. This manual approach not only slowed down the process but compromised the accuracy of our RPN values and overall risk reliability.
The Technical Solution: The Pwark Engine
I engineered Pwark to replace this manual chain with a single, validated digital workflow:
To solve these challenges, I built a centralized "Source of Truth" by migrating plant data into a Python-managed SQLite database. I developed a custom extraction tool to bridge the gap between legacy Word-based FMEAs and our new digital system, turning static documents into actionable data. Finally, I automated the RPN calculation process, allowing the system to instantly flag high-risk failures the moment data is entered.
The Practical Result
Engineered a Python-based automation system to centralize complex FMEA data silos, achieving a 99% reduction in reporting errors and a 40% increase in operational efficiency through real-time risk tracking.
The Quantitative Connection
Building Pwark taught me that in complex systems, the highest risks often come from manual data handling.
My research focused on identifying the Bullwhip Effect—the phenomenon where small changes in consumer demand lead to large, inefficient inventory swings across a supply chain.
The Technical Approach
The Impact