Audio-Text-to-Text
Transformers
Safetensors
Chinese
English
qwen2_5_omni_thinker
speech
audio
speech-evaluation
expressive-speech
mandarin
chain-of-thought
ceaeval
Instructions to use TianRW/CEAEval-Model with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use TianRW/CEAEval-Model with Transformers:
# Load model directly from transformers import AutoTokenizer, GatedAttenQwen2_5omnithinker tokenizer = AutoTokenizer.from_pretrained("TianRW/CEAEval-Model") model = GatedAttenQwen2_5omnithinker.from_pretrained("TianRW/CEAEval-Model") - Notebooks
- Google Colab
- Kaggle
| {% set audio_count = namespace(value=0) %}{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system | |
| You are a helpful assistant.<|im_end|> | |
| {% endif %}<|im_start|>{{ message['role'] }} | |
| {% if message['content'] is string %}{{ message['content'] }}<|im_end|> | |
| {% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_bos|><|IMAGE|><|vision_eos|>{% elif content['type'] == 'audio' or 'audio' in content or 'audio_url' in content %}{% set audio_count.value = audio_count.value + 1 %}{% if add_audio_id %}Audio {{ audio_count.value }}: {% endif %}<|audio_bos|><|AUDIO|><|audio_eos|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_bos|><|VIDEO|><|vision_eos|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|> | |
| {% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant | |
| {% endif %} |