You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A high-performance API server that provides OpenAI-compatible endpoints for MLX models. Developed using Python and powered by the FastAPI framework, it provides an efficient, scalable, and user-friendly solution for running MLX-based vision and language models locally with an OpenAI-compatible interface.
Defensive layer for mlx / mlx-lm / mlx-vlm on Apple Silicon. Prevents IOGPUFamily kernel panics, SIGABRT from Metal completion handlers, and Mac reboots from MLX inference. Covers 7 known kernel-panic root-cause classes.
macMLX brings local LLM inference to Apple Silicon with a first-class native macOS experience. No cloud, no telemetry, no Electron — just your Mac running models at full speed.
Streamlit app using Google's MedGemma 1.5 4B (via mlx-vlm on Apple Silicon) for chest X-ray analysis: anatomy localization with bounding boxes and longitudinal comparison of two CXRs.