diff --git a/.gitattributes b/.gitattributes index 0cec2425e9d79f7f0bbf28b0c229a2b6b166c6ac..1d106ec848dda05696abd065e59c1d09fe637cc0 100644 --- a/.gitattributes +++ b/.gitattributes @@ -51,3 +51,4 @@ examples/SQA/SLUE-P2-SQA5-Test/sample_2.wav filter=lfs diff=lfs merge=lfs -text examples/SQA/Spoken-Squad-v1/sample_0.wav filter=lfs diff=lfs merge=lfs -text examples/SQA/Spoken-Squad-v1/sample_1.wav filter=lfs diff=lfs merge=lfs -text examples/SQA/Spoken-Squad-v1/sample_2.wav filter=lfs diff=lfs merge=lfs -text +examples/SQA/DREAM-TTS-MCQ-Test/sample_2.wav filter=lfs diff=lfs merge=lfs -text diff --git a/examples/AC/AudioCaps-Test/dataset_info.json b/examples/AC/AudioCaps-Test/dataset_info.json index 2b8cfe680340090d702b3d6b88a8378e4b9449fc..82148686a795bb258e6676260855fb8cf9ef19e4 100644 --- a/examples/AC/AudioCaps-Test/dataset_info.json +++ b/examples/AC/AudioCaps-Test/dataset_info.json @@ -118,6 +118,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/AC/AudioCaps-Test/sample_0.wav b/examples/AC/AudioCaps-Test/sample_0.wav index 4331131a6543acb7d4f6e69a71708518f65bc034..e37f34b9f07a7b6266957992824bab914ce15fe3 100644 Binary files a/examples/AC/AudioCaps-Test/sample_0.wav and b/examples/AC/AudioCaps-Test/sample_0.wav differ diff --git a/examples/AC/AudioCaps-Test/sample_1.wav b/examples/AC/AudioCaps-Test/sample_1.wav index 2fdb280d6d9a2912c0aad84bc423a3b17ab4acc7..9f16faee9805c2cbf530d36958c99f82952f42f3 100644 Binary files a/examples/AC/AudioCaps-Test/sample_1.wav and b/examples/AC/AudioCaps-Test/sample_1.wav differ diff --git a/examples/AC/AudioCaps-Test/sample_2.wav b/examples/AC/AudioCaps-Test/sample_2.wav index bd06790c5bba36aa0c69693e84c4bb40cdfbc290..95c13a1a0c7e8fce564fba1e7ecd506797ef83ed 100644 Binary files a/examples/AC/AudioCaps-Test/sample_2.wav and b/examples/AC/AudioCaps-Test/sample_2.wav differ diff --git a/examples/AC/AudioCaps-Test/state.json b/examples/AC/AudioCaps-Test/state.json index 65cc6f6b6acacba97b729c23d8639556e02847f6..b334a469898f2333372cff0d2c4e4def008d271e 100644 --- a/examples/AC/AudioCaps-Test/state.json +++ b/examples/AC/AudioCaps-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "0e301916c3676d35", + "_fingerprint": "e654a4081bc1365b", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/AC/WavCaps-Test/dataset_info.json b/examples/AC/WavCaps-Test/dataset_info.json index 0854513dfb2583899f6f4539b14e6b2982daf86b..acf11db0c4cc1869f7763270ddadbfe4c30f73d4 100644 --- a/examples/AC/WavCaps-Test/dataset_info.json +++ b/examples/AC/WavCaps-Test/dataset_info.json @@ -114,6 +114,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/AC/WavCaps-Test/sample_0.wav b/examples/AC/WavCaps-Test/sample_0.wav index a76589f0898c5033ab99b0a15f34564fcd4e0d89..86803de3706292dd68f1a49c6dc5b66662eabae8 100644 Binary files a/examples/AC/WavCaps-Test/sample_0.wav and b/examples/AC/WavCaps-Test/sample_0.wav differ diff --git a/examples/AC/WavCaps-Test/sample_1.wav b/examples/AC/WavCaps-Test/sample_1.wav index 011a74daa10980002a95da355e3534e42f69bf6c..00182dc3b08c94349036bf2f6da3df783e7358dc 100644 Binary files a/examples/AC/WavCaps-Test/sample_1.wav and b/examples/AC/WavCaps-Test/sample_1.wav differ diff --git a/examples/AC/WavCaps-Test/sample_2.wav b/examples/AC/WavCaps-Test/sample_2.wav index d8afaa47b29c26dd49fbc93a95e3d8310a402a11..a9c4f74921215f29b6e4e01b3ec87b98d4c64a04 100644 Binary files a/examples/AC/WavCaps-Test/sample_2.wav and b/examples/AC/WavCaps-Test/sample_2.wav differ diff --git a/examples/AC/WavCaps-Test/state.json b/examples/AC/WavCaps-Test/state.json index 1157a81d2f767520f45c0be2fe9bab4ff33f96b2..18c89274b45b312d6a38f3f6e978483c100b9db1 100644 --- a/examples/AC/WavCaps-Test/state.json +++ b/examples/AC/WavCaps-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "1e570096603c2a32", + "_fingerprint": "ce408e4cfa3eec8a", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Common-Voice-15-En-Test/dataset_info.json b/examples/ASR/Common-Voice-15-En-Test/dataset_info.json index 0ee97efe251f1c21a87ef3c09372d7249c0d657c..61c4c1050499fd3220dbfe6013858b008c8d9810 100644 --- a/examples/ASR/Common-Voice-15-En-Test/dataset_info.json +++ b/examples/ASR/Common-Voice-15-En-Test/dataset_info.json @@ -34,11 +34,11 @@ }, "other_attributes": { "accents": { - "dtype": "null", + "dtype": "string", "_type": "Value" }, "age": { - "dtype": "null", + "dtype": "string", "_type": "Value" }, "client_id": { @@ -50,7 +50,7 @@ "_type": "Value" }, "gender": { - "dtype": "null", + "dtype": "string", "_type": "Value" }, "language": { @@ -146,6 +146,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Common-Voice-15-En-Test/sample_0.wav b/examples/ASR/Common-Voice-15-En-Test/sample_0.wav index 0037bc557881e74bdcc687e1d28af287a6c22344..98ad00ab81a597db93fb42634a7b173662c5402e 100644 Binary files a/examples/ASR/Common-Voice-15-En-Test/sample_0.wav and b/examples/ASR/Common-Voice-15-En-Test/sample_0.wav differ diff --git a/examples/ASR/Common-Voice-15-En-Test/sample_1.wav b/examples/ASR/Common-Voice-15-En-Test/sample_1.wav index bc3c6afd5fe31546b091231f643b37f05ba11595..2ff5b62f10f2614553c20a23a94c67dd8813f7c7 100644 Binary files a/examples/ASR/Common-Voice-15-En-Test/sample_1.wav and b/examples/ASR/Common-Voice-15-En-Test/sample_1.wav differ diff --git a/examples/ASR/Common-Voice-15-En-Test/sample_2.wav b/examples/ASR/Common-Voice-15-En-Test/sample_2.wav index 4245350f9951353f48a156651b2b0bc3629ae84a..98eff0b7069351279a995d5e33aa12b83a109dd0 100644 Binary files a/examples/ASR/Common-Voice-15-En-Test/sample_2.wav and b/examples/ASR/Common-Voice-15-En-Test/sample_2.wav differ diff --git a/examples/ASR/Common-Voice-15-En-Test/state.json b/examples/ASR/Common-Voice-15-En-Test/state.json index 0dc2de30ddac4cca90d5224840099629052ffe14..3064aa54130945bc783ab898011d0e84385e02fa 100644 --- a/examples/ASR/Common-Voice-15-En-Test/state.json +++ b/examples/ASR/Common-Voice-15-En-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "30218d56801da2e8", + "_fingerprint": "6342d438049fbc7e", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Earnings21-Test/dataset_info.json b/examples/ASR/Earnings21-Test/dataset_info.json index 08b3d046e3630b0d47e267cf866bc1ce79106af7..1bf27e3aaa89f2fa43812252ac2377fab8ae1708 100644 --- a/examples/ASR/Earnings21-Test/dataset_info.json +++ b/examples/ASR/Earnings21-Test/dataset_info.json @@ -110,6 +110,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Earnings21-Test/state.json b/examples/ASR/Earnings21-Test/state.json index 206bda8429e14fa71c1fbe9ddf26f730e0359b97..37a0a357127feb8eb1497306a502edb53a1e4eee 100644 --- a/examples/ASR/Earnings21-Test/state.json +++ b/examples/ASR/Earnings21-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "2e3dea299b387757", + "_fingerprint": "d0ad1703cbc51418", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Earnings22-Test/dataset_info.json b/examples/ASR/Earnings22-Test/dataset_info.json index 08b3d046e3630b0d47e267cf866bc1ce79106af7..1bf27e3aaa89f2fa43812252ac2377fab8ae1708 100644 --- a/examples/ASR/Earnings22-Test/dataset_info.json +++ b/examples/ASR/Earnings22-Test/dataset_info.json @@ -110,6 +110,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Earnings22-Test/state.json b/examples/ASR/Earnings22-Test/state.json index 643700cb9278c071e7a0ca46b2760cf18034ad11..ad7966db6aa47d1e1dbff1f49d710fa0e2050484 100644 --- a/examples/ASR/Earnings22-Test/state.json +++ b/examples/ASR/Earnings22-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "c2ddf91e8ccb230c", + "_fingerprint": "fb047ff90ed3a443", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/GigaSpeech-Test/dataset_info.json b/examples/ASR/GigaSpeech-Test/dataset_info.json index 7056dd861bf6cf034028f9f01c5153ac3e4ec304..1dd0025578e934e74b979da9e81789eedd9a2f29 100644 --- a/examples/ASR/GigaSpeech-Test/dataset_info.json +++ b/examples/ASR/GigaSpeech-Test/dataset_info.json @@ -138,6 +138,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/GigaSpeech-Test/sample_0.wav b/examples/ASR/GigaSpeech-Test/sample_0.wav index f2afdc748f9c58eb2e6cd9793b8a642471d23c31..073483a0e4604323d0369ff2216b2a5765780823 100644 Binary files a/examples/ASR/GigaSpeech-Test/sample_0.wav and b/examples/ASR/GigaSpeech-Test/sample_0.wav differ diff --git a/examples/ASR/GigaSpeech-Test/sample_1.wav b/examples/ASR/GigaSpeech-Test/sample_1.wav index 942b01315d851d0ed1a72ae561e6859c9fbb32fd..ca1b14f3ed3ea483c331f9acbaabde5b03c0b35b 100644 Binary files a/examples/ASR/GigaSpeech-Test/sample_1.wav and b/examples/ASR/GigaSpeech-Test/sample_1.wav differ diff --git a/examples/ASR/GigaSpeech-Test/sample_2.wav b/examples/ASR/GigaSpeech-Test/sample_2.wav index 0e61bad88fd0422f1f69e07c1d79da115df77c7e..11094ecca41233498c7b38e51b9b70971f113ff6 100644 Binary files a/examples/ASR/GigaSpeech-Test/sample_2.wav and b/examples/ASR/GigaSpeech-Test/sample_2.wav differ diff --git a/examples/ASR/GigaSpeech-Test/state.json b/examples/ASR/GigaSpeech-Test/state.json index e61469ffa3f7414712f1f7fa7fa6525b49e5cac4..dd846b8228bc8c6a5d6d88f7300ff35a6b51e5cb 100644 --- a/examples/ASR/GigaSpeech-Test/state.json +++ b/examples/ASR/GigaSpeech-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "0032f92a85e94025", + "_fingerprint": "84a02614da440215", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/IMDA-Part1-ASR-Test/sample_0.wav b/examples/ASR/IMDA-Part1-ASR-Test/sample_0.wav index 9f1b0abb84dc9bb30a04f2b08eb0f0be0ec33cf4..ea36e7a02a6771ea96e210107b81ef3a5e5cd791 100644 Binary files a/examples/ASR/IMDA-Part1-ASR-Test/sample_0.wav and b/examples/ASR/IMDA-Part1-ASR-Test/sample_0.wav differ diff --git a/examples/ASR/IMDA-Part1-ASR-Test/sample_1.wav b/examples/ASR/IMDA-Part1-ASR-Test/sample_1.wav index 2f5b49a926a9509ad34700c9d00df8bbda49fd57..553aa3569726cd67a40baf3f50c7b5e18f32cd74 100644 Binary files a/examples/ASR/IMDA-Part1-ASR-Test/sample_1.wav and b/examples/ASR/IMDA-Part1-ASR-Test/sample_1.wav differ diff --git a/examples/ASR/IMDA-Part1-ASR-Test/sample_2.wav b/examples/ASR/IMDA-Part1-ASR-Test/sample_2.wav index f36eb9efe06c44ea5cbd63d26257e8d37f0079b7..e4a9780515675dbb05b95cff26fb5aea26fe9aa7 100644 Binary files a/examples/ASR/IMDA-Part1-ASR-Test/sample_2.wav and b/examples/ASR/IMDA-Part1-ASR-Test/sample_2.wav differ diff --git a/examples/ASR/IMDA-Part1-ASR-Test/state.json b/examples/ASR/IMDA-Part1-ASR-Test/state.json index 6e63ed10109432151f62c147e30d2704343a756d..d8919f5bfc220aed30d07a397179f30478ac3176 100644 --- a/examples/ASR/IMDA-Part1-ASR-Test/state.json +++ b/examples/ASR/IMDA-Part1-ASR-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "9b9c1437475afa9d", + "_fingerprint": "1514e693988caee7", "_format_columns": [ "context", "instruction", diff --git a/examples/ASR/IMDA-Part2-ASR-Test/sample_0.wav b/examples/ASR/IMDA-Part2-ASR-Test/sample_0.wav index fe0422bcb8356543fbde4cc9981779ca712dcc22..1784a0b78a165892ba01586b04974a15dda4eea3 100644 Binary files a/examples/ASR/IMDA-Part2-ASR-Test/sample_0.wav and b/examples/ASR/IMDA-Part2-ASR-Test/sample_0.wav differ diff --git a/examples/ASR/IMDA-Part2-ASR-Test/sample_1.wav b/examples/ASR/IMDA-Part2-ASR-Test/sample_1.wav index 4f6ec7d97638278d3cf05791b44d2d072ab161ba..1fde75aea28f1557cd2fe3d80434d8b394f116ac 100644 Binary files a/examples/ASR/IMDA-Part2-ASR-Test/sample_1.wav and b/examples/ASR/IMDA-Part2-ASR-Test/sample_1.wav differ diff --git a/examples/ASR/IMDA-Part2-ASR-Test/sample_2.wav b/examples/ASR/IMDA-Part2-ASR-Test/sample_2.wav index d1e955f3a2f6d402e2661703d4880a4881c088c8..260cd5f87122a500ebf3b014bb2c323cc1200abe 100644 Binary files a/examples/ASR/IMDA-Part2-ASR-Test/sample_2.wav and b/examples/ASR/IMDA-Part2-ASR-Test/sample_2.wav differ diff --git a/examples/ASR/IMDA-Part2-ASR-Test/state.json b/examples/ASR/IMDA-Part2-ASR-Test/state.json index 2dbed7b83dd6665624b87d3b11c514bc97e22f29..21299bb3dfc3a9f29854e5a39c5e2130dfca6bae 100644 --- a/examples/ASR/IMDA-Part2-ASR-Test/state.json +++ b/examples/ASR/IMDA-Part2-ASR-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "1b048a2e54d0c002", + "_fingerprint": "8a8e117080f24a8b", "_format_columns": [ "context", "instruction", diff --git a/examples/ASR/LibriSpeech-Test-Clean/dataset_info.json b/examples/ASR/LibriSpeech-Test-Clean/dataset_info.json index b7936842ee323be9e4c2360418e44a417cf0f626..5a91f71cfb2044e6060c8f395ee4b798384d32d8 100644 --- a/examples/ASR/LibriSpeech-Test-Clean/dataset_info.json +++ b/examples/ASR/LibriSpeech-Test-Clean/dataset_info.json @@ -122,6 +122,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/LibriSpeech-Test-Clean/sample_0.wav b/examples/ASR/LibriSpeech-Test-Clean/sample_0.wav index 266ab3ddafe3c5d192606c6cdf579238b0150e62..2a103bec37460090a764f280ebcf4791ccd17d4e 100644 Binary files a/examples/ASR/LibriSpeech-Test-Clean/sample_0.wav and b/examples/ASR/LibriSpeech-Test-Clean/sample_0.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Clean/sample_1.wav b/examples/ASR/LibriSpeech-Test-Clean/sample_1.wav index 238f6c95cf91c6bff49636351557961b3e06d356..60af90099e82c9115285de4897e38412a9440cac 100644 Binary files a/examples/ASR/LibriSpeech-Test-Clean/sample_1.wav and b/examples/ASR/LibriSpeech-Test-Clean/sample_1.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Clean/sample_2.wav b/examples/ASR/LibriSpeech-Test-Clean/sample_2.wav index f127981ac0a35d93e4fcf3607a84858e9ea276d9..d44035f1bacdce219e44056683944fd9cfecf7eb 100644 Binary files a/examples/ASR/LibriSpeech-Test-Clean/sample_2.wav and b/examples/ASR/LibriSpeech-Test-Clean/sample_2.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Clean/state.json b/examples/ASR/LibriSpeech-Test-Clean/state.json index 1c1c2c843a1f7da02ed60cbeb427026c76972e7d..103a2443a4c9e5355798ee842efb63b6bf286af8 100644 --- a/examples/ASR/LibriSpeech-Test-Clean/state.json +++ b/examples/ASR/LibriSpeech-Test-Clean/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "c56d3af03a1dc565", + "_fingerprint": "cb0a09e53f0cc5db", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/LibriSpeech-Test-Other/dataset_info.json b/examples/ASR/LibriSpeech-Test-Other/dataset_info.json index b7936842ee323be9e4c2360418e44a417cf0f626..5a91f71cfb2044e6060c8f395ee4b798384d32d8 100644 --- a/examples/ASR/LibriSpeech-Test-Other/dataset_info.json +++ b/examples/ASR/LibriSpeech-Test-Other/dataset_info.json @@ -122,6 +122,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/LibriSpeech-Test-Other/sample_0.wav b/examples/ASR/LibriSpeech-Test-Other/sample_0.wav index 9ceae84addeaa1cfd86a5ae164ba3f48c7994c2f..595f7ba137749fd1578688365357e230351f6660 100644 Binary files a/examples/ASR/LibriSpeech-Test-Other/sample_0.wav and b/examples/ASR/LibriSpeech-Test-Other/sample_0.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Other/sample_1.wav b/examples/ASR/LibriSpeech-Test-Other/sample_1.wav index 5facbda4871665edf9aefcd3e58b6056b296c233..948927ee05b269f31a635b2682751b4f0d50a8c5 100644 Binary files a/examples/ASR/LibriSpeech-Test-Other/sample_1.wav and b/examples/ASR/LibriSpeech-Test-Other/sample_1.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Other/sample_2.wav b/examples/ASR/LibriSpeech-Test-Other/sample_2.wav index 237b56f851d3989319b0fe06ce85878ab69c4850..82b4b2afd731d0c2a1501e34c5689cff431955a0 100644 Binary files a/examples/ASR/LibriSpeech-Test-Other/sample_2.wav and b/examples/ASR/LibriSpeech-Test-Other/sample_2.wav differ diff --git a/examples/ASR/LibriSpeech-Test-Other/state.json b/examples/ASR/LibriSpeech-Test-Other/state.json index cd23da8c954dfb155a1fd333adb7cfbdfcde8eaf..f296f217907538df40fe8adfabb15d59e5f8d326 100644 --- a/examples/ASR/LibriSpeech-Test-Other/state.json +++ b/examples/ASR/LibriSpeech-Test-Other/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "5fc28a37097fe19f", + "_fingerprint": "0ed80d8ca27350ce", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Peoples-Speech-Test/dataset_info.json b/examples/ASR/Peoples-Speech-Test/dataset_info.json index 0a6ab198f1ae81f599b4abcef59dba0043ab934e..a8d496de62c861269b2a5c8bf9826fedc8abf807 100644 --- a/examples/ASR/Peoples-Speech-Test/dataset_info.json +++ b/examples/ASR/Peoples-Speech-Test/dataset_info.json @@ -114,6 +114,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Peoples-Speech-Test/sample_0.wav b/examples/ASR/Peoples-Speech-Test/sample_0.wav index 560d3df564e6658b9d7040aff205a47737e16dd2..d18ce818679d916f7d285ab41c4eb0b4c64ab80a 100644 Binary files a/examples/ASR/Peoples-Speech-Test/sample_0.wav and b/examples/ASR/Peoples-Speech-Test/sample_0.wav differ diff --git a/examples/ASR/Peoples-Speech-Test/sample_1.wav b/examples/ASR/Peoples-Speech-Test/sample_1.wav index 1513c2587700588fda3c13c14f037bb821dfd6df..7b6d8264d363621884f39814ea63380cb64434a4 100644 Binary files a/examples/ASR/Peoples-Speech-Test/sample_1.wav and b/examples/ASR/Peoples-Speech-Test/sample_1.wav differ diff --git a/examples/ASR/Peoples-Speech-Test/sample_2.wav b/examples/ASR/Peoples-Speech-Test/sample_2.wav index 8ab7248ab6278f1ad098d67fee87fec1a934b0bc..ec498dbdb1c5d0bac49a03779da1f995bd621c07 100644 Binary files a/examples/ASR/Peoples-Speech-Test/sample_2.wav and b/examples/ASR/Peoples-Speech-Test/sample_2.wav differ diff --git a/examples/ASR/Peoples-Speech-Test/state.json b/examples/ASR/Peoples-Speech-Test/state.json index 37d3ef70f3164ef13c8bd4d576e7d8b4eb9a4511..41de0c42892f63c7ec1efb30e3d62a0be9f877a8 100644 --- a/examples/ASR/Peoples-Speech-Test/state.json +++ b/examples/ASR/Peoples-Speech-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "6203edc47e9a3c56", + "_fingerprint": "bf71989dac1baa0c", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Tedlium3-Longform-Test/dataset_info.json b/examples/ASR/Tedlium3-Longform-Test/dataset_info.json index ce0eff747512d4783e49763b36beac50f103dce5..cf9965f85a972806830fe14f38cc250366f94118 100644 --- a/examples/ASR/Tedlium3-Longform-Test/dataset_info.json +++ b/examples/ASR/Tedlium3-Longform-Test/dataset_info.json @@ -122,6 +122,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Tedlium3-Longform-Test/state.json b/examples/ASR/Tedlium3-Longform-Test/state.json index a1041e8ced6bf9491bb3d91d8d0f8d5814c98494..acb7115eb5d31d5d7a8e9c066448ea7e12bed918 100644 --- a/examples/ASR/Tedlium3-Longform-Test/state.json +++ b/examples/ASR/Tedlium3-Longform-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "f89ae31db7413bf4", + "_fingerprint": "16ff45685ee4694b", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/ASR/Tedlium3-Test/dataset_info.json b/examples/ASR/Tedlium3-Test/dataset_info.json index ce0eff747512d4783e49763b36beac50f103dce5..cf9965f85a972806830fe14f38cc250366f94118 100644 --- a/examples/ASR/Tedlium3-Test/dataset_info.json +++ b/examples/ASR/Tedlium3-Test/dataset_info.json @@ -122,6 +122,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/ASR/Tedlium3-Test/sample_0.wav b/examples/ASR/Tedlium3-Test/sample_0.wav index c7bd8c3b68ed10585921e11bee7d24a34339f78c..94362f231d21545289d87680f8580d7dd2de71e8 100644 Binary files a/examples/ASR/Tedlium3-Test/sample_0.wav and b/examples/ASR/Tedlium3-Test/sample_0.wav differ diff --git a/examples/ASR/Tedlium3-Test/sample_1.wav b/examples/ASR/Tedlium3-Test/sample_1.wav index 3b684a181c9fd4ac1d95c2fc7988379a26f677cb..8a73e7a2a9ec7b50f172450c75d9d16a2b7c0987 100644 Binary files a/examples/ASR/Tedlium3-Test/sample_1.wav and b/examples/ASR/Tedlium3-Test/sample_1.wav differ diff --git a/examples/ASR/Tedlium3-Test/sample_2.wav b/examples/ASR/Tedlium3-Test/sample_2.wav index b1ab681404201295b8fb1355c1c29d92147a750e..79460a0b295f5b0ee22515fcbb7fcdee5ce83816 100644 Binary files a/examples/ASR/Tedlium3-Test/sample_2.wav and b/examples/ASR/Tedlium3-Test/sample_2.wav differ diff --git a/examples/ASR/Tedlium3-Test/state.json b/examples/ASR/Tedlium3-Test/state.json index c0eec8964db7c884c56d55d6cae98406a6c52de4..5db8543827fbeb809394b23b9163137b6d9f7b71 100644 --- a/examples/ASR/Tedlium3-Test/state.json +++ b/examples/ASR/Tedlium3-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "fde82500f706458b", + "_fingerprint": "33116bb8d5c8b7bf", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/SI/ALPACA-Audio-Test/dataset_info.json b/examples/SI/ALPACA-Audio-Test/dataset_info.json index 7dd927db4ab336b38a99b686aced46f2e4d9a5cb..882789dac9ac95ae77467581f55b4f296746e7a9 100644 --- a/examples/SI/ALPACA-Audio-Test/dataset_info.json +++ b/examples/SI/ALPACA-Audio-Test/dataset_info.json @@ -126,6 +126,28 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "audio_text_instruction": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/SI/ALPACA-Audio-Test/sample_0.wav b/examples/SI/ALPACA-Audio-Test/sample_0.wav index 3d516af829afc1afcea07064b2aaaec734175e61..9d59769e693ba8186a468c920bf33aed6415b9a0 100644 Binary files a/examples/SI/ALPACA-Audio-Test/sample_0.wav and b/examples/SI/ALPACA-Audio-Test/sample_0.wav differ diff --git a/examples/SI/ALPACA-Audio-Test/sample_1.wav b/examples/SI/ALPACA-Audio-Test/sample_1.wav index 3707bfb5a929acff874947ad6ea76b902ddbc2bf..9fa77dc3a558295c617e6ba5f9c77d4a744f18fe 100644 Binary files a/examples/SI/ALPACA-Audio-Test/sample_1.wav and b/examples/SI/ALPACA-Audio-Test/sample_1.wav differ diff --git a/examples/SI/ALPACA-Audio-Test/sample_2.wav b/examples/SI/ALPACA-Audio-Test/sample_2.wav index 70248c9fd2ff3b5792099536311d05bed53fba47..e7446eb105a62c5223dc799407a78a303ffa3273 100644 Binary files a/examples/SI/ALPACA-Audio-Test/sample_2.wav and b/examples/SI/ALPACA-Audio-Test/sample_2.wav differ diff --git a/examples/SI/ALPACA-Audio-Test/state.json b/examples/SI/ALPACA-Audio-Test/state.json index cba82888d784fef2f53a04b162e712eb4813e7dd..3b1309df321667faea7ee972e3d92df95efd575f 100644 --- a/examples/SI/ALPACA-Audio-Test/state.json +++ b/examples/SI/ALPACA-Audio-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "b48e6d66699d09cc", + "_fingerprint": "c6a29ee9e25361a7", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/SI/OpenHermes-Audio-Test/sample_0.wav b/examples/SI/OpenHermes-Audio-Test/sample_0.wav index 3a11a8017ed68f552b088939360efbd713fb23d3..58c30ae68669cf5651f7dcd57e28e73b9347f677 100644 Binary files a/examples/SI/OpenHermes-Audio-Test/sample_0.wav and b/examples/SI/OpenHermes-Audio-Test/sample_0.wav differ diff --git a/examples/SI/OpenHermes-Audio-Test/sample_1.wav b/examples/SI/OpenHermes-Audio-Test/sample_1.wav index c5bb85768cc4a7c6f7206fc9c56412a614d08d9a..58ffc5f5c796a545f6ec10417f286ad13acc64cd 100644 Binary files a/examples/SI/OpenHermes-Audio-Test/sample_1.wav and b/examples/SI/OpenHermes-Audio-Test/sample_1.wav differ diff --git a/examples/SI/OpenHermes-Audio-Test/sample_2.wav b/examples/SI/OpenHermes-Audio-Test/sample_2.wav index 63ae3d47c286000d5b4180e65a0aa12437dbe7d6..7d013e06ccb1a7b66fa1371cd7dd25a0067c6424 100644 Binary files a/examples/SI/OpenHermes-Audio-Test/sample_2.wav and b/examples/SI/OpenHermes-Audio-Test/sample_2.wav differ diff --git a/examples/SI/OpenHermes-Audio-Test/state.json b/examples/SI/OpenHermes-Audio-Test/state.json index 7dd94d080c16d3c9654a1ba57c8142e0b9c1380d..83993faa18ef244ce7cd82d1a262742737e1b751 100644 --- a/examples/SI/OpenHermes-Audio-Test/state.json +++ b/examples/SI/OpenHermes-Audio-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "2edb418fca8ac348", + "_fingerprint": "44d9e5a3612e0df7", "_format_columns": [ "context", "instruction", diff --git a/examples/SQA/CN-College-Listen-MCQ-Test/dataset_info.json b/examples/SQA/CN-College-Listen-MCQ-Test/dataset_info.json index c3526b447e28b1c62fece39889011d58962d741f..81903da2dc3bf5b849827427d5a61f932fe485bb 100644 --- a/examples/SQA/CN-College-Listen-MCQ-Test/dataset_info.json +++ b/examples/SQA/CN-College-Listen-MCQ-Test/dataset_info.json @@ -45,78 +45,6 @@ "dtype": "string", "_type": "Value" } - }, - "salmonn_7b": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "wavllm_fairseq": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "whisper_large_v3_with_llama_3_8b_instruct": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "qwen_audio_chat": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } } }, "homepage": "", diff --git a/examples/SQA/CN-College-Listen-MCQ-Test/sample_0.wav b/examples/SQA/CN-College-Listen-MCQ-Test/sample_0.wav index 6923b4c01dd221f14c21f85df1e8165eb5b59bbf..b383286458adcd302192c1a8840bf8adf43b1c41 100644 Binary files a/examples/SQA/CN-College-Listen-MCQ-Test/sample_0.wav and b/examples/SQA/CN-College-Listen-MCQ-Test/sample_0.wav differ diff --git a/examples/SQA/CN-College-Listen-MCQ-Test/sample_1.wav b/examples/SQA/CN-College-Listen-MCQ-Test/sample_1.wav index 750f1b8414060167651ac33128408b510b1545ca..fcdf3095b8c9b36fd8d1140fabc9c22e506ce290 100644 Binary files a/examples/SQA/CN-College-Listen-MCQ-Test/sample_1.wav and b/examples/SQA/CN-College-Listen-MCQ-Test/sample_1.wav differ diff --git a/examples/SQA/CN-College-Listen-MCQ-Test/sample_2.wav b/examples/SQA/CN-College-Listen-MCQ-Test/sample_2.wav index 0a7aa93de8bade69d708bbf36a8954843b02c2cb..3674d26ff67ffee9fbf6cd64bd00af74906a37d0 100644 Binary files a/examples/SQA/CN-College-Listen-MCQ-Test/sample_2.wav and b/examples/SQA/CN-College-Listen-MCQ-Test/sample_2.wav differ diff --git a/examples/SQA/CN-College-Listen-MCQ-Test/state.json b/examples/SQA/CN-College-Listen-MCQ-Test/state.json index 75ad396b15b37427ae3dc61c702735f6de828994..c6cc629f7c25521767ac482dbd1e71307ff9c443 100644 --- a/examples/SQA/CN-College-Listen-MCQ-Test/state.json +++ b/examples/SQA/CN-College-Listen-MCQ-Test/state.json @@ -4,16 +4,12 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "a9d1aad6ff31612e", + "_fingerprint": "efbbc5b180ee96e0", "_format_columns": [ "context", "instruction", "answer", - "other_attributes", - "salmonn_7b", - "wavllm_fairseq", - "whisper_large_v3_with_llama_3_8b_instruct", - "qwen_audio_chat" + "other_attributes" ], "_format_kwargs": {}, "_format_type": null, diff --git a/examples/SQA/DREAM-TTS-MCQ-Test/dataset_info.json b/examples/SQA/DREAM-TTS-MCQ-Test/dataset_info.json index c4eadac83d7d01b0964e29d70f224e2d0a3a9246..a16938258fabe5c46b36019db75cd53e5132c21f 100644 --- a/examples/SQA/DREAM-TTS-MCQ-Test/dataset_info.json +++ b/examples/SQA/DREAM-TTS-MCQ-Test/dataset_info.json @@ -59,78 +59,6 @@ "dtype": "string", "_type": "Value" } - }, - "salmonn_7b": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "wavllm_fairseq": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "whisper_large_v3_with_llama_3_8b_instruct": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } - }, - "qwen_audio_chat": { - "answer": { - "dtype": "string", - "_type": "Value" - }, - "model_prediction": { - "dtype": "string", - "_type": "Value" - }, - "task_type": { - "dtype": "string", - "_type": "Value" - }, - "text": { - "dtype": "string", - "_type": "Value" - } } }, "homepage": "", diff --git a/examples/SQA/DREAM-TTS-MCQ-Test/sample_1.wav b/examples/SQA/DREAM-TTS-MCQ-Test/sample_1.wav index 36cb70eb458d7650de9d8fce1d8b46dac6ca53f2..8f6275ca50ba72f0800b5392777d63d5d1655a2f 100644 Binary files a/examples/SQA/DREAM-TTS-MCQ-Test/sample_1.wav and b/examples/SQA/DREAM-TTS-MCQ-Test/sample_1.wav differ diff --git a/examples/SQA/DREAM-TTS-MCQ-Test/state.json b/examples/SQA/DREAM-TTS-MCQ-Test/state.json index 70faed9032446a81505aa07f10cc97f424fd3026..9839f8b445e8be16c0eacbf29b36a184b7e09cf0 100644 --- a/examples/SQA/DREAM-TTS-MCQ-Test/state.json +++ b/examples/SQA/DREAM-TTS-MCQ-Test/state.json @@ -4,16 +4,12 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "e5f8ea158b4a9ca3", + "_fingerprint": "1414881d04766772", "_format_columns": [ "answer", "context", "instruction", - "other_attributes", - "salmonn_7b", - "wavllm_fairseq", - "whisper_large_v3_with_llama_3_8b_instruct", - "qwen_audio_chat" + "other_attributes" ], "_format_kwargs": {}, "_format_type": null, diff --git a/examples/SQA/Public-SG-Speech-QA-Test/sample_2.wav b/examples/SQA/Public-SG-Speech-QA-Test/sample_2.wav index 19bef4b0a13042e08e5c59f0bdde09868c33bff1..9911e47fed2232a4c5cd141221b8e8bc98ed66cf 100644 Binary files a/examples/SQA/Public-SG-Speech-QA-Test/sample_2.wav and b/examples/SQA/Public-SG-Speech-QA-Test/sample_2.wav differ diff --git a/examples/SQA/Public-SG-Speech-QA-Test/state.json b/examples/SQA/Public-SG-Speech-QA-Test/state.json index 579a4a84b69f43df3d260b1e814859aa751a0126..cb60981a9b1ca32983da76e04178d039a7507568 100644 --- a/examples/SQA/Public-SG-Speech-QA-Test/state.json +++ b/examples/SQA/Public-SG-Speech-QA-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "77292393d596f5fa", + "_fingerprint": "794fdda22fd93c7f", "_format_columns": [ "context", "instruction", diff --git a/examples/SQA/SLUE-P2-SQA5-Test/dataset_info.json b/examples/SQA/SLUE-P2-SQA5-Test/dataset_info.json index d0f327fa13416dff7d7cc45e7f9652d37ea2b296..6c681f17c89b019564bbd2d0509865a05ce4db3a 100644 --- a/examples/SQA/SLUE-P2-SQA5-Test/dataset_info.json +++ b/examples/SQA/SLUE-P2-SQA5-Test/dataset_info.json @@ -155,6 +155,24 @@ "_type": "Value" } }, + "mowe_audio": { + "answer": { + "dtype": "string", + "_type": "Value" + }, + "model_prediction": { + "dtype": "string", + "_type": "Value" + }, + "task_type": { + "dtype": "string", + "_type": "Value" + }, + "text": { + "dtype": "string", + "_type": "Value" + } + }, "qwen_audio_chat": { "answer": { "dtype": "string", diff --git a/examples/SQA/SLUE-P2-SQA5-Test/state.json b/examples/SQA/SLUE-P2-SQA5-Test/state.json index 5c88181b585af5561ec3cfbd92c9d80a9e2f1f57..2c3aad4c97f82ceb8147cd280eaa55e3d858b29d 100644 --- a/examples/SQA/SLUE-P2-SQA5-Test/state.json +++ b/examples/SQA/SLUE-P2-SQA5-Test/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "35d15f00aac76485", + "_fingerprint": "8a250cf20de7599a", "_format_columns": [ "context", "instruction", @@ -14,6 +14,7 @@ "wavllm_fairseq", "Qwen2-Audio-7B-Instruct", "whisper_large_v3_with_llama_3_8b_instruct", + "mowe_audio", "qwen_audio_chat" ], "_format_kwargs": {}, diff --git a/examples/SQA/Spoken-Squad-v1/state.json b/examples/SQA/Spoken-Squad-v1/state.json index 89840561a23f1218e34bb60af320658effae2d2c..64ffce20f87758120e69473703b377a002af8178 100644 --- a/examples/SQA/Spoken-Squad-v1/state.json +++ b/examples/SQA/Spoken-Squad-v1/state.json @@ -4,7 +4,7 @@ "filename": "data-00000-of-00001.arrow" } ], - "_fingerprint": "1d8937f6900ae52c", + "_fingerprint": "d63972b6ec648239", "_format_columns": [ "context", "instruction",