Neural Magic
Neural Magic (Acquired by Red Hat) empowers developers to optimize & deploy LLMs at scale. Our model compression & acceleration enable top performance with vLLM
Pinned Loading
Repositories
Showing 10 of 89 repositories
- mini-swe-agent Public Forked from SWE-agent/mini-swe-agent
The 100 line AI agent that solves GitHub issues or helps you in your command line. Radically simple, no huge configs, no giant monorepo—but scores >74% on SWE-bench verified!
neuralmagic/mini-swe-agent’s past year of commit activity - nyann_poker Public
neuralmagic/nyann_poker’s past year of commit activity - model-validation-configs Public
neuralmagic/model-validation-configs’s past year of commit activity - GuardBench Public Forked from eldarkurtic/GuardBench
A Python library for guardrail models evaluation with vLLM support.
neuralmagic/GuardBench’s past year of commit activity - sglang Public Forked from sgl-project/sglang
SGLang is a fast serving framework for large language models and vision language models.
neuralmagic/sglang’s past year of commit activity - vllm-fork Public Forked from tlrmchlsmth/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
neuralmagic/vllm-fork’s past year of commit activity - tpu-inference Public Forked from vllm-project/tpu-inference
TPU inference for vLLM, with unified JAX and PyTorch support.
neuralmagic/tpu-inference’s past year of commit activity - DeepEP Public Forked from deepseek-ai/DeepEP
DeepEP: an efficient expert-parallel communication library
neuralmagic/DeepEP’s past year of commit activity
Top languages
Loading…