Skip to content
View back2matching's full-sized avatar

Block or report back2matching

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
back2matching/README.md
back2matching



     


$ cat ~/.stack
Python       ████████████░░░░░░░░  59%
TypeScript   █████░░░░░░░░░░░░░░░  27%
JavaScript   ██░░░░░░░░░░░░░░░░░░   8%

+ React · Next.js · Node · Tailwind · Supabase · Redis · Docker · Solidity

> now

  • FlockRun · AI agent runtime, getting them to actually cooperate
  • cigoL · reverse logic engine, thinks backwards on purpose
  • matching.work · the portfolio, next.js + gsap, full brutalist


code. ship. repeat.

Popular repositories Loading

  1. turboquant turboquant Public

    First open-source TurboQuant KV cache compression for LLM inference. Drop-in for HuggingFace. pip install turboquant.

    Python 19 5

  2. back2matching back2matching Public

  3. kvcache-bench kvcache-bench Public

    Benchmark every KV cache compression method on your GPU. One command, real numbers. Supports Ollama + llama.cpp.

    Python

  4. quant-sim quant-sim Public

    Which quantization should I use? One command benchmarks every quant level on YOUR GPU.

    Python

  5. turboquant-vectors turboquant-vectors Public

    Compress embeddings 6x instantly with TurboQuant. First pip package using Google's TurboQuant (ICLR 2026) for vector search. 71.9% recall vs FAISS PQ 13.3%.

    Python