Skip to content
View RamseyvanderMeer's full-sized avatar

Highlights

  • Pro

Organizations

@ieee-uiuc

Block or report RamseyvanderMeer

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
RamseyvanderMeer/README.md

Ramsey van der Meer

I build AI systems that actually have to run in production: RAG apps, agentic workflows, evaluation tools, auth-heavy backend services, and the infrastructure around them.

Most of my recent work sits at the edge of AI and systems engineering. I like taking a rough model-powered idea and turning it into something observable, secure, and reliable enough for real users.

What I Work On

  • AI applications with LLMs, RAG, embeddings, vector search, and evaluation loops
  • Backend systems in Python, FastAPI, Go, TypeScript, PostgreSQL, Redis, and Kafka
  • Production infrastructure with GCP, Docker, Kubernetes, Terraform, GitHub Actions, Prometheus, and Grafana
  • Agentic workflows, browser automation, data pipelines, and reliability tooling

Featured Projects

An AI-powered automotive diagnostics platform that turns fault codes, vehicle context, and technical repair content into repair guidance. Built with FastAPI, conversational RAG, ChromaDB, GCP Cloud Run, Google OAuth, Redis rate limits, and secure API access.

Demo: mist-expo.vercel.app/sign-in

A Grok + LlamaIndex research agent over simulated X data. It plans, decomposes tasks, calls tools, retrieves context with hybrid semantic/keyword search, replans when results are weak, and exports evaluation metrics.

No hosted demo right now because it depends on model API keys, but the repo includes Docker setup, mock data generation, and evaluation outputs so the workflow can be run locally.

A lightweight human-preference ranking app for comparing LLM responses. It uses pairwise voting and ELO-style updates to turn subjective model outputs into a more useful ranking signal.

An early evaluation pipeline for comparing LLM-generated data corrections across OpenAI, Anthropic, Gemini, Llama, and Mistral-style providers, with JSON repair and scoring workflows.

No hosted demo; this is mainly a reproducible evaluation/prototyping repo for comparing provider behavior and failure modes.

Links

Pinned Loading

  1. mist mist Public

    AI automotive diagnostics platform using FastAPI, RAG, ChromaDB, GCP Cloud Run, OAuth, and feedback-driven retrieval.

    Python

  2. social-media-inference-bot social-media-inference-bot Public

    Autonomous Grok/LlamaIndex research agent with hybrid retrieval, ChromaDB vector search, replanning, context compression, and evaluation metrics.

    Python

  3. ieee ieee Public

    Full-stack IEEE UIUC website rebuild with event content management, member profiles, attendance logging, and REST APIs.

    TypeScript

  4. llm-testing llm-testing Public

    LLM evaluation sandbox for structured data correction, JSON repair, provider comparison, and model-as-judge scoring.

    Jupyter Notebook

  5. rankingElo rankingElo Public

    Human-preference ranking app for comparing LLM responses with pairwise voting and ELO-style scoring.

    JavaScript