Slimmed, cleaned and fine-tuned oh-my-opencode fork, consumes much less tokens
-
Updated
Apr 24, 2026 - TypeScript
Slimmed, cleaned and fine-tuned oh-my-opencode fork, consumes much less tokens
🦖 X—LLM: Cutting Edge & Easy LLM Finetuning
Production-grade RAG API built in Rust. Hybrid search with HNSW dense vectors and BM25 sparse matching, cross-encoder reranking, layout-aware document extraction via Docling, and 94.5% accuracy on Open RAG Bench. Powered by Cerebras, Groq, Milvus, and Jina AI.
A robust Node.js proxy server that automatically rotates API keys for Gemini and OpenAI APIs when rate limits (429 errors) are encountered. Built with zero dependencies and comprehensive logging.
PHP library for interacting with AI platform provider.
OpenAI-compatible gateway aggregating 8 free LLM providers with automatic failover, circuit breakers, and smart routing. 32+ models, zero cost.
Ultra-fast, customizable AI voice dictation in any active app on Windows (MacOS and Linux coming soon)
A tool keep tabs on your Cerebras Code usage limits, in real time
AI-powered geopolitical news intelligence platform. Ingests 100K+ daily events from GDELT, stores in MotherDuck (DuckDB), orchestrates with Dagster, and features an AI chat interface with Text-to-SQL. Full data engineering stack at $0/month.
🦖 X—LLM: Simple & Cutting Edge LLM Finetuning
This repository features an example of how to utilize the xllm library. Included is a solution for a common type of assessment given to LLM engineers
Matrix decomposition and multiplication on the Cerebras Wafer-Scale Engine (WSE) architecture
A solution that could prioritize patients based on urgency, reducing wait times and ensuring those who need immediate care.
Add a description, image, and links to the cerebras topic page so that developers can more easily learn about it.
To associate your repository with the cerebras topic, visit your repo's landing page and select "manage topics."