Tools for understanding how transformer predictions are built layer-by-layer
-
Updated
Jun 2, 2024 - Python
Tools for understanding how transformer predictions are built layer-by-layer
This paper list focuses on the theoretical and empirical analysis of language models, especially large language models (LLMs). The papers in this list investigate the learning behavior, generalization ability, and other properties of language models through theoretical analysis, empirical analysis, or a combination of both.
Lumina-T2X is a unified framework for Text to Any Modality Generation
Accelerate your training with this open-source library. Optimize performance with streamlined training and serving options with JAX. 🚀
Implementation of Alphafold 3 in Pytorch
This project utilizes the power of BERT (Bidirectional Encoder Representations from Transformers) for sentiment analysis
Ongoing research training transformer models at scale
🎤📄 An innovative tool that transforms audio or video files into text transcripts and generates concise meeting minutes. Stay organized and efficient in your meetings, and get ready for Phase 2 where we'll be open for contributions to enable real-time meeting transcription! 🚀
A low-memory high-performance CPU-based API for Meta's No Language Left Behind (NLLB) using CTranslate2, hosted on Hugging Face Spaces.
AI research lab🔬: implementations of AI papers and theoretical research: InstructGPT, llama, transformers, diffusion models, RLHF, etc...
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Testing the capabilities of the Llama 3 language model, specifically the Meta-Llama-3-8B-Instruct variant with 8 billion parameters.
LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.
Genshin Impact Character Chat Models tuned by Lora on LLM
⚡️SwanLab: your ML experiment notebook. 你的AI实验笔记本,跟踪与可视化你的机器学习全流程
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
Official implementation for "UniST: A Prompt-Empowered Universal Model for Urban Spatio-Temporal Prediction" (KDD 2024)
Context aware, pluggable and customizable data protection and de-identification SDK for text and images
🔮 SuperDuperDB: Bring AI to your database! Build, deploy and manage any AI application directly with your existing data infrastructure, without moving your data. Including streaming inference, scalable model training and vector search.
Add a description, image, and links to the transformers topic page so that developers can more easily learn about it.
To associate your repository with the transformers topic, visit your repo's landing page and select "manage topics."