A high-throughput and memory-efficient inference and serving engine for LLMs
-
Updated
Jun 12, 2024 - Python
A high-throughput and memory-efficient inference and serving engine for LLMs
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Large Language Model Text Generation Inference
🔧 A Kotlin coroutine wrapper around Media3's Transformer API.
Official implementation of "CST-YOLO: A Novel Method for Blood Cell Detection Based on Improved YOLOv7 and CNN-Swin Transformer".
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
A paper list of some recent Transformer-based CV works.
A framework for few-shot evaluation of language models.
A Python-based toolkit for comparing transformers.
My Bachelor's Thesis
A powerful HTTP client for Dart and Flutter, which supports global settings, Interceptors, FormData, aborting and canceling a request, files uploading and downloading, requests timeout, custom adapters, etc.
An offline CPU-first memory-scarce chat application to perform RAG on your corpus of data. Powered by OpenChat and CTranslate2.
An attention based approach to convert Indian Sign Language to Text using simulated hand gesture data
A toolbox of vision models and algorithms based on MindSpore
[CVPR 2024] Official Implementation of Collaborating Foundation models for Domain Generalized Semantic Segmentation
Port of OpenAI's Whisper model in C/C++
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Add a description, image, and links to the transformer topic page so that developers can more easily learn about it.
To associate your repository with the transformer topic, visit your repo's landing page and select "manage topics."