Architecture Benchmarks – Review & Extension#1176
Open
Conversation
…rchitecture-adapter
…chitecture-adapter
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Re-ran all benchmarks against a collection of models (
gpt2,neo,pythia,OPT,Qwen2,Bloom,OpenELM), and resolved any new discrepancies caused by updated to transformers v5, and all of our other latest changes.Added new Text Quality benchmark, which runs a generation of text and scores it with GPT2 to ensure that we are generating valid human-readable text
Stabilizing float types – some benchmark comparisons were converting back and forth between float16 and float32 due to the source type of the model. These updates stabilize those types to allow for more accurate testing of the model's accuracy when loaded via
TransformerBridgeResolved bugs discovered in
bloom_attentionResolved deprecation issues caused by transformers v5 in T5
the following models were failing to properly generate or pass the other benchmarks, updated Architecture adapter to pass base benchmarks & new generate benchmark
Cleaned up any duplicate code or unused benchmark functions that have been superseded by newer, better testing
Bug fix (non-breaking change which fixes an issue)
New feature (non-breaking change which adds functionality)
Checklist: