@@ -728,7 +728,7 @@ def test_resize_embeddings_untied_with_deepspeed(self):
728
728
@require_torch
729
729
class Qwen2_5OmniModelIntegrationTest (unittest .TestCase ):
730
730
def setUp (self ):
731
- self .processor = AutoProcessor .from_pretrained ("/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test " )
731
+ self .processor = AutoProcessor .from_pretrained ("Qwen/Qwen2.5-Omni-7B " )
732
732
self .audio_url = "https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2-Audio/audio/glass-breaking-151256.mp3"
733
733
self .audio_url_additional = (
734
734
"https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2-Audio/audio/f2641_0_throatclearing.wav"
@@ -758,9 +758,7 @@ def tearDown(self):
758
758
759
759
@slow
760
760
def test_small_model_integration_test (self ):
761
- model = Qwen2_5OmniModel .from_pretrained (
762
- "/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test" , torch_dtype = torch .float32 , device_map = "auto"
763
- )
761
+ model = Qwen2_5OmniModel .from_pretrained ("Qwen/Qwen2.5-Omni-7B" , torch_dtype = torch .float32 , device_map = "auto" )
764
762
765
763
text = self .processor .apply_chat_template (self .messages , tokenize = False , add_generation_prompt = True )
766
764
inputs = self .processor (
@@ -818,9 +816,7 @@ def test_small_model_integration_test(self):
818
816
819
817
@slow
820
818
def test_small_model_integration_test_batch (self ):
821
- model = Qwen2_5OmniModel .from_pretrained (
822
- "/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test" , torch_dtype = torch .float32 , device_map = "auto"
823
- )
819
+ model = Qwen2_5OmniModel .from_pretrained ("Qwen/Qwen2.5-Omni-7B" , torch_dtype = torch .float32 , device_map = "auto" )
824
820
text = self .processor .apply_chat_template (self .messages , tokenize = False , add_generation_prompt = True )
825
821
inputs = self .processor (
826
822
text = [text , text ],
@@ -844,9 +840,7 @@ def test_small_model_integration_test_batch(self):
844
840
845
841
@slow
846
842
def test_small_model_integration_test_multiturn (self ):
847
- model = Qwen2_5OmniModel .from_pretrained (
848
- "/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test" , torch_dtype = torch .float32 , device_map = "auto"
849
- )
843
+ model = Qwen2_5OmniModel .from_pretrained ("Qwen/Qwen2.5-Omni-7B" , torch_dtype = torch .float32 , device_map = "auto" )
850
844
851
845
messages = [
852
846
self .messages [0 ],
@@ -883,9 +877,7 @@ def test_small_model_integration_test_multiturn(self):
883
877
884
878
@slow
885
879
def test_small_model_integration_test_w_audio (self ):
886
- model = Qwen2_5OmniModel .from_pretrained (
887
- "/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test" , torch_dtype = torch .float32 , device_map = "auto"
888
- )
880
+ model = Qwen2_5OmniModel .from_pretrained ("Qwen/Qwen2.5-Omni-7B" , torch_dtype = torch .float32 , device_map = "auto" )
889
881
audio_url = "https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2-Audio/audio/guess_age_gender.wav"
890
882
891
883
messages = [
@@ -918,7 +910,7 @@ def test_small_model_integration_test_w_audio(self):
918
910
@require_torch_gpu
919
911
def test_small_model_integration_test_batch_flashatt2 (self ):
920
912
model = Qwen2_5OmniModel .from_pretrained (
921
- "/home/xj_data/lvyuanjun.lyj/res/navit_thinker_talker/test " ,
913
+ "Qwen/Qwen2.5-Omni-7B " ,
922
914
torch_dtype = torch .bfloat16 ,
923
915
attn_implementation = "flash_attention_2" ,
924
916
device_map = "auto" ,
0 commit comments