Marium Ashraf profile picture

Hello, I'm

Marium Ashraf

|

My LinkedIn profile My Github profile

Get To Know More

About Me

Profile picture
Experience icon

Experience

3+ years
AI Systems, LLM Engineering & Full-Stack Development

Education icon

Education

B.Eng. Computer Engineering (Major in Software Engineering)
Toronto Metropolitan University
Graduating April 2027

I'm an Operations & AI Engineer building production-grade systems that combine LLM intelligence with distributed architecture. At Lyrata, I design AI-assisted tooling that reduces operational latency by 90% and cuts onboarding time in half. My work spans microservice LLM evaluation pipelines, vector database systems, and full-stack applications — all built with observability, fault tolerance, and scalability in mind.

From GPT-4o-powered RAG systems and MLflow experiment tracking to containerized microservices and real-time analytics dashboards, I build systems that teams can trust at scale. I've improved model quality by 35%, reduced debugging cycles by 60%, and accelerated feature delivery by 30% through architectural refactors and standardized workflows.

Arrow icon

Key

Achievements & Impact

0
%

SOP Lookup Latency Reduction

AI-assisted tooling at Lyrata

📊
0
+

Model Outputs Processed Daily

Scalable evaluation pipelines

🚀
0
%

Onboarding Time Reduction

LyrataGPT SDK-like interface

Arrow icon

Explore My

Experience

Operations & AI Intern, Lyrata

July 2025 – Present

Designed and deployed AI-assisted tooling to parse, validate, and surface operational knowledge from structured and unstructured sources (Python, Flask, PostgreSQL, ChromaDB, GPT-4o), reducing SOP lookup latency by ~90% and enabling consistent, auditable system behavior. Built a production-grade internal "SDK-like" interface (LyrataGPT) that delivers cited, procedure-accurate responses to technicians, improving developer-style ergonomics for non-technical users and cutting onboarding ramp time by ~50%. Applied Vector Institute DaRMoD methodologies to produce ML systems, implementing reproducible evaluation pipelines, MLflow experiment tracking, containerized services, and fault-tolerant routing under real-world conditions.

Prompt & AI Engineer, Outlier

November 2022 – July 2025

Developed microservice LLM evaluation pipelines with automated rubric scoring, anomaly detection, and usage analytics, scaling to daily inferences while improving model quality by ~35%. Built SQL-backed observability and analytics systems with real-time dashboards and automated alerts, reducing debugging cycles by ~60% and improving reliability of production deployments. Led architectural refactors of distributed inference infrastructure, optimizing container orchestration, caching strategies, and load balancing to improve latency and system stability under heavy traffic. Translated business and product requirements into technical specifications for cross-functional teams, introducing standardized evaluation and release workflows that accelerated feature delivery by ~30%.

Web Developer Intern, While She is True

January 2022 – December 2023

Built production full-stack features using React and backend APIs, including authentication flows, SQL data models, and Stripe integrations, ensuring secure and reliable client-server communication. Worked against clearly defined API contracts, contributing to stable interface boundaries between frontend and backend systems and reducing integration errors during feature iteration. Implemented automated testing and CI/CD pipelines (Jest, React Testing Library, GitHub Actions) to support predictable releases, minimize regressions, and improve system reliability over time.

Arrow icon

Explore My

Skills & Technologies

Languages & AI/LLMs

Languages: Python, Java, JavaScript/TypeScript

AI & LLMs: GPT-4o, LLM evaluation, RAG, embeddings, semantic analysis, MLflow. I build retrieval-augmented generation systems using vector databases, design automated evaluation pipelines that process thousands of model outputs daily, and implement MLflow experiment tracking for reproducible ML workflows. My work has improved model quality by 35% through systematic evaluation and optimization.

Distributed Systems & APIs

Distributed Systems: Docker, microservices, message queues (RabbitMQ), observability, fault tolerance

APIs & SDKs: REST, gRPC, Slack API, SDK design patterns, versioning. I architect containerized microservices with fault-tolerant routing, implement observability stacks with real-time dashboards, and design SDK-like interfaces that improve developer ergonomics. My distributed automation systems integrate Slack APIs, Google Sheets, and databases with schema versioning and idempotent handlers.

Databases, Cloud & Testing

Databases: PostgreSQL, MongoDB, BigQuery, vector databases (ChromaDB)

Cloud & DevOps: AWS, GCP (Pub/Sub, Airflow), CI/CD, GitHub Actions, Linux

Data & ML Systems: ETL/ELT, analytics pipelines, feature engineering, usage telemetry

Testing & QA: PyTest, Jest, automated test suites, validation pipelines. I build SQL-backed observability systems, implement comprehensive CI/CD pipelines, and develop automated testing frameworks that reduce regressions and improve system reliability.

Arrow icon

Browse My Recent

Projects

Hugging Face Sentiment API preview

Hugging Face Sentiment API

🐍 🐳 🤗 🚀
Low-Latency Async I/O Scalable

A containerized FastAPI inference service integrating Hugging Face Transformers for real-time text sentiment prediction. Designed with Uvicorn for async concurrency and validated via Jupyter-based parallel request testing.

Built with Docker and the distilbert-base-uncased-finetuned-sst-2-english model for optimized low-latency inference. Features async I/O, batching, and scaling via Gunicorn workers with future-ready plans for NGINX load balancing, structured logging, and Prometheus metrics.

FastAPI Observability Demo preview

FastAPI Observability Demo

🐍 📊 🔍 🐳 📡
Full Observability Load Tested Production-Ready

A FastAPI microservice demonstrating end-to-end observability through Grafana, Prometheus, Tempo, and Loki. Features OpenTelemetry instrumentation for trace-to-log correlation and metric exemplars, deployed via Docker Compose.

Implements trace injection across three FastAPI services using the OpenTelemetry SDK with Prometheus metrics, Tempo tracing, and Loki log pipelines. Load tested via Locust and k6 to validate distributed tracing and resource efficiency, with metrics visualized in Grafana dashboards for real-time debugging and performance monitoring.

Personal Finance Dashboard preview

Personal Finance Dashboard

🐍 📈 🗄️ 🐳 🐼
ETL Pipeline Data Visualization Containerized

A Streamlit-based analytics app for visualizing Bluecoins expense data using Python, SQL, and Docker. Designed to support intuitive budget tracking and category insights with a responsive, minimal interface.

Developed an ETL pipeline from CSV → SQL → Streamlit UI using pandas and SQLAlchemy, containerized via Docker Compose for reproducibility. Provides dynamic charts on income trends, category flow, and net balance with future integrations planned for OAuth login and hosted deployment.

Robo Advisor preview

Robo Advisor

☁️ 💬 🧠 🚀
Serverless NLP-Powered AWS Architecture

A serverless financial chatbot that guides users through retirement investment planning. Built with AWS Lex for NLP dialogue management and AWS Lambda for real-time portfolio recommendations based on user risk profiles.

Implements Lambda-backed validation for user intents and secure financial calculations with algorithm-driven portfolio allocation based on risk tolerance and horizon. Built with serverless architecture for scalability and low-cost deployment, featuring an integrated conversational flow designed for accuracy, responsiveness, and ease of use.

Arrow icon

Get in Touch

Contact Me

Let's Connect

I'm always interested in discussing new opportunities, collaborations, or just having a chat about technology and innovation.