Enhance model card for GlimpsePrune: Add metadata and comprehensive details
#1
by nielsr HF Staff - opened
This PR significantly enhances the model card for GlimpsePrune by:
- Adding
pipeline_tag: image-text-to-text: This improves discoverability on the Hugging Face Hub, allowing users to find the model under relevant Vision-Language tasks (e.g., athttps://huggingface.co/models?pipeline_tag=image-text-to-text). - Specifying
library_name: transformers: This indicates compatibility with the Hugging Face Transformers library, enabling the "Use in Transformers" snippet button and facilitating easy integration for users. - Linking to the paper: The model card now includes a direct link to the paper "A Glimpse to Compress: Dynamic Visual Token Pruning for Large Vision-Language Models" on Hugging Face Papers.
- Including the GitHub repository link: A link to the official code repository
https://github.com/HVision-NKU/GlimpsePruneis now prominent. - Adding detailed model description: Key information from the paper abstract and the project's GitHub README, including a general overview, key features, framework overview, and performance results, has been integrated. This provides a clear understanding of the model's purpose, capabilities, and technical details.
- Providing a clear usage example: A Python code snippet demonstrating how to use the model with the Hugging Face
transformerslibrary for inference is included, making it easier for users to get started. - Embedding illustrative images: Key images from the project's GitHub repository are embedded to visually explain the framework and showcase results.
- Including citation information: The BibTeX entry for the paper is added for proper attribution.
This PR aims to make the GlimpsePrune model more accessible and informative for the community.
ashun989 changed pull request status to merged