File size: 3,543 Bytes
734e414
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
roles_map = {
    'system': 'system',
    'user': 'user',
    'human': 'user',
    'assistant': 'assistant',
    'gpt': 'assistant',
    'AI': 'assistant',
}

R1_SYSTEM_PROMPT = '''\
You are an AI assistant.

Your primary directive is to provide well-reasoned, structured, and extensively detailed responses.

Formatting Requirements:
- Always structure your replies using: <think>{reasoning}</think>{answer}
- The <think></think> block should contain at least six reasoning steps when applicable.
- If the answer requires minimal thought, the <think></think> block may be left empty.
- The user does not see the <think></think> section. Any information critical to the response must be included in the answer.
- If you notice that you have engaged in circular reasoning or repetition, immediately terminate {reasoning} with a </think> and proceed to the {answer}

Response Guidelines:
- Detailed and Structured: Use rich Markdown formatting for clarity and readability.
- Scientific and Logical Approach: Your explanations should reflect the depth and precision of the greatest scientific minds.
- Prioritize Reasoning: Always reason through the problem first, unless the answer is trivial.
- Concise yet Complete: Ensure responses are informative, yet to the point without unnecessary elaboration.
- Maintain a professional, intelligent, and analytical tone in all interactions.'''

core_reason_datasets = [
    #
    # math reason
    #
    # 8.43 GB, 450,258
    *[
        {'kind': 'instruct', 'path': 'open-r1/OpenR1-Math-220k', 'data_dir': 'data', 'split': f'train[{i}%:{i + 10}%]', 'field': 'messages', 'transform': lambda msgs: [
            {'role': roles_map[m['from']], 'content': m['value']}
            for m in msgs
        ]}
        for i in range(0, 100, 10)
    ],

    #
    # general reason
    #
    # 3.55 GB, 227,914
    *[
        {'kind': 'instruct', 'path': 'open-thoughts/OpenThoughts-114k', 'data_dir': 'data', 'split': f'train[{i}%:{i + 10}%]', 'transform': lambda r: [
            {'role': 'system', 'content': r['system']}
        ] + [
            {'role': roles_map[m['from']], 'content': m['value']}
            for m in r['conversations']
        ]}
        for i in range(0, 100, 10)
    ],
    # 3.98 GB, 814,334
    # 300k
    *[
        {'kind': 'instruct', 'path': 'cognitivecomputations/dolphin-r1', 'data_files': 'dolphin-r1-reasoning-deepseek.jsonl', 'split': f'train[{i}%:{i + 10}%]', 'transform': lambda r: [
            {'role': 'system', 'content': R1_SYSTEM_PROMPT},
            *r['messages'],
            {'role': 'assistant', 'content': '<think>\n' + (r.get('reasoning') or '') + '\n</think>\n' + (r.get('answer') or '')},
        ]}
        for i in range(0, 100, 10)
    ],
    # 300k
    *[
        {'kind': 'instruct', 'path': 'cognitivecomputations/dolphin-r1', 'data_files': 'dolphin-r1-reasoning-flash.jsonl', 'split': f'train[{i}%:{i + 10}%]', 'transform': lambda r: [
            {'role': 'system', 'content': R1_SYSTEM_PROMPT},
            *r['messages'],
            {'role': 'assistant', 'content': '<think>\n' + (r.get('reasoning') or '') + '\n</think>\n' + (r.get('answer') or '')},
        ]}
        for i in range(0, 100, 10)
    ],
    # 21.1 MB, 1,000
    {'kind': 'instruct', 'path': 'simplescaling/s1K-1.1', 'split': 'train', 'transform': lambda r: [
        {'role': 'user', 'content': r.get('question') or ''},
        {'role': 'assistant', 'content': '<think>\n' + (r.get('deepseek_thinking_trajectory') or '') + '\n</think>\n' + (r.get('solution') or '')},
    ]}
]