A GPU-based correlator for MeerKAT Extension
-
Updated
Jun 12, 2024 - Python
A GPU-based correlator for MeerKAT Extension
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, Axolotl, etc.
A high-performance inference system for large language models, designed for production environments.
✨ Zero-code distributed tracing and profiling, observability via eBPF 🚀
Open deep learning compiler stack for cpu, gpu and specialized accelerators
AMD ROCm Performance Primitives (RPP) library is a comprehensive high-performance computer vision library for AMD processors with HIP/OpenCL/CPU back-ends.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
☁️ VRAM for SDXL, AnimateDiff, and upscalers. Run your workflows on the cloud, from your local ComfyUI
Monte Carlo eXtreme for OpenCL (MCXCL)
A Pythonic framework to simplify AI service building
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
CUDA C++ Core Libraries
Add a description, image, and links to the gpu topic page so that developers can more easily learn about it.
To associate your repository with the gpu topic, visit your repo's landing page and select "manage topics."