Welcome
MLOps / LLMOps Engineer at LynxMind Portugal · PhD Researcher at FEUP · Junior Researcher at INESC TEC · Building production-grade AI systems for research and industry
About
I work as an MLOps and LLMOps Engineer at LynxMind Portugal, where I have been building and operating production-grade AI systems for more than three years. My work focuses on cloud-native platforms, scalable ML/LLM architectures, deployment automation, monitoring, and continuous evaluation for real operating environments.
I hold a BSc in Computer Science from Universidade Católica de Moçambique and an MSc in Applied Informatics from Universidade de Aveiro, developed in collaboration with Aily Labs in Barcelona. I am currently a PhD Researcher in Electrical and Computer Engineering at FEUP and a Junior Researcher at INESC TEC within LIAAD.
My trajectory combines academic research and applied engineering across Europe and Asia, including China, with experience in multicultural and distributed teams. I am particularly interested in reproducible AI systems, MLOps observability, LLM evaluation, and reliable delivery of intelligent platforms.
My training also includes Deep Learning from Stanford University, Generative AI specialization at FCUP, and IBM Ethical Hacker certification, complementing my engineering background with strong foundations in modern AI methods and secure system design.
Download my CV
Design and operation of ML and LLM systems across experimentation, orchestration, evaluation, serving, and lifecycle governance for production environments.
Scalable cloud-native platforms for AI and engineering workloads, with infrastructure as code, container orchestration, and resilient deployment patterns.
Continuous integration and delivery pipelines for applications and AI platforms, with automated validation, promotion strategies, and release reliability.
Applied AI systems for research and industry, combining data pipelines, feature engineering, model development, experimentation, and operational delivery.
Monitoring, telemetry, traceability, and security-aware operations for AI and cloud workloads, improving reliability, visibility, and controlled delivery.
Production operations, troubleshooting, systems administration, and environment support for reliable services and engineering teams.
MLOps Observability, LLM Evaluation, Reproducible AI, AI Systems Engineering, production reliability for intelligent systems, and cloud-native platforms for research and industry.
3+ years delivering production AI systems in industry, combining research, deployment automation, observability, and engineering workflows across academic and operational environments.
Built a production-grade retrieval and generation platform for enterprise knowledge workflows, with ingestion, evaluation, response tracing, and deployment-ready serving patterns.
Technologies: LangChain, LlamaIndex, FastAPI, Vector Search, Docker, Kubernetes
Impact: Structured GenAI delivery around reproducibility, observability, and controlled rollout.
Designed a robust agent-based workflow for task decomposition, tool use, prompt routing, and response validation in complex AI-assisted processes.
Technologies: LangChain, Prompt Engineering, FastAPI, Python, Evaluation Pipelines
Impact: Advanced prototyping and operationalization of complex Generative AI workflows.
Built an end-to-end MLOps platform for training, tracking, versioning, serving, and release control across model lifecycle stages.
Technologies: Kubeflow, MLflow, Airflow, Docker, Kubernetes, Terraform
Impact: Reduced manual model promotion steps and improved consistency across delivery environments.
Implemented monitoring, tracing, and evaluation layers for model and platform behavior, with visibility into performance, drift signals, and operational health.
Technologies: Prometheus, Grafana, OpenTelemetry, MLflow, Python
Impact: Improved troubleshooting speed and observability maturity in AI operations.
Designed a GitOps-oriented deployment model with environment promotion, automated validation, rollback readiness, and version-controlled infrastructure changes.
Technologies: GitHub Actions, Kubernetes, Terraform, Bash, Docker
Impact: Strengthened release governance and deployment traceability.
Engineered a cloud-native platform for application and AI workloads with reusable CI/CD components, infrastructure automation, and environment standardization.
Technologies: AWS, Azure, Docker, Kubernetes, GitHub Actions, Terraform
Impact: Increased delivery consistency and reduced configuration drift.
Combined secure delivery practices with runtime visibility, hardening, and operational controls for services deployed in cloud-native environments.
Technologies: Linux, CI/CD, Monitoring, Security Controls, Observability Tooling
Impact: Improved deployment confidence and operational resilience.
Email: leonel.silima0@gmail.com