-
Notifications
You must be signed in to change notification settings - Fork 35
Issues: explosion/curated-transformers
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Truncation of sequences that are beyond the model's maximum length
feat/tokenization
Feature: Tokenization/piecer
type/bug
Type: Bug
type/feature
Type: Feature
#359
opened Jan 14, 2024 by
MootezSaaD
Add suggested PyTorch LLM optimizations
feat/generation
Feature: Generation
feat/model
Feature: models
#356
opened Dec 1, 2023 by
danieldk
Add support for attention sinks
feat/layers
Feature: Layers
feat/model
Feature: models
type/feature
Type: Feature
Expose more outputs through the Feature: Generation
type/feature
Type: Feature
Generator
interface
feat/generation
Convert QKV projection splitting methods into Torch modules
feat/layers
Feature: Layers
type/maintenance
Type: Maintenance
Option to only return the last hidden layer output from models
feat/model
Feature: models
type/feature
Type: Feature
Support for Encoder-Decoder-style architectures
feat/model
Feature: models
type/feature
Type: Feature
Add Low-Rank Adapters injection into base models
feat/training
Feature: Training/Fine-tuning
type/feature
Type: Feature
ProTip!
Mix and match filters to narrow down what you’re looking for.