fix capital letters
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ library_name: transformers
|
|
19 |
</p>
|
20 |
|
21 |
## What's New
|
22 |
-
- [2025.06.06] **MiniCPM4** series are released! This model achieves ultimate efficiency improvements while maintaining optimal performance at the same scale! It can achieve over 5x generation acceleration on typical end-side chips! You can find technical report [here](https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf).🔥🔥🔥
|
23 |
- [2025.06.09] **MiniCPM4-8B-mlx** is available and you can run MiniCPM4 on your Apple devices! Thanks to [pzc163](https://huggingface.co/pzc163) for providing this converted model version and related usage instructions.
|
24 |
|
25 |
## MiniCPM4 Series
|
@@ -61,14 +61,14 @@ MiniCPM 4 is an extremely efficient edge-side large model that has undergone eff
|
|
61 |
Here is a guide on how to run the `MiniCPM-8B-mlx` model from the command line using `mlx-lm`. You can use mlx-lm to interact with the `MiniCPM-8B-mlx` model directly from your command line. This is a powerful tool that allows you to quickly test and use LLMs in the MLX format.
|
62 |
|
63 |
### Basic Usage
|
64 |
-
Here is a specific example. This command will load the `
|
65 |
```Bash
|
66 |
-
mlx_lm.generate --model
|
67 |
```
|
68 |
|
69 |
### MLX-LM Command Line Parameters
|
70 |
- `mlx_lm.generate`: This is the primary command in the mlx-lm toolkit used for text generation.
|
71 |
-
- `--model
|
72 |
- `--prompt "..."`: This parameter is used to provide the initial text that you want the model to respond to or complete.
|
73 |
- `--max-tokens`: Sets the maximum number of tokens to generate. For example, `--max-tokens 200` will limit the output to 200 tokens.
|
74 |
- `--temp`: Controls the randomness of the output. Higher temperature values (like 0.8) will produce more diverse and creative outputs, while lower values (like 0.2) will make the output more deterministic and focused. The default value is usually 0.6.
|
@@ -78,7 +78,7 @@ mlx_lm.generate --model OpenBMB/MiniCPM-8B-mlx --prompt "hello, pls tell me whic
|
|
78 |
|
79 |
The following command will use a higher temperature value and limit the output length:
|
80 |
```bash
|
81 |
-
mlx_lm.generate --model
|
82 |
--prompt "tell me a story about a robot who discovered music" \
|
83 |
--max-tokens 500 \
|
84 |
--temp 0.8
|
|
|
19 |
</p>
|
20 |
|
21 |
## What's New
|
22 |
+
- [2025.06.06] **MiniCPM4** series are released! This model achieves ultimate efficiency improvements while maintaining optimal performance at the same scale! It can achieve over 5x generation acceleration on typical end-side chips! You can find the technical report [here](https://github.com/OpenBMB/MiniCPM/tree/main/report/MiniCPM_4_Technical_Report.pdf).🔥🔥🔥
|
23 |
- [2025.06.09] **MiniCPM4-8B-mlx** is available and you can run MiniCPM4 on your Apple devices! Thanks to [pzc163](https://huggingface.co/pzc163) for providing this converted model version and related usage instructions.
|
24 |
|
25 |
## MiniCPM4 Series
|
|
|
61 |
Here is a guide on how to run the `MiniCPM-8B-mlx` model from the command line using `mlx-lm`. You can use mlx-lm to interact with the `MiniCPM-8B-mlx` model directly from your command line. This is a powerful tool that allows you to quickly test and use LLMs in the MLX format.
|
62 |
|
63 |
### Basic Usage
|
64 |
+
Here is a specific example. This command will load the `openbmb/MiniCPM-8B-mlx` model and generate text based on the prompt you provide: "hello, pls tell me which one is the most powerful LLM in the World".
|
65 |
```Bash
|
66 |
+
mlx_lm.generate --model openbmb/MiniCPM-8B-mlx --prompt "hello, pls tell me which one is the most powerful LLM in the World"
|
67 |
```
|
68 |
|
69 |
### MLX-LM Command Line Parameters
|
70 |
- `mlx_lm.generate`: This is the primary command in the mlx-lm toolkit used for text generation.
|
71 |
+
- `--model openbmb/MiniCPM-8B-mlx`: This parameter specifies the model to be loaded. `openbmb/MiniCPM-8B-mlx` is the model's identifier on the Hugging Face Hub. mlx-lm will automatically download and cache the model from there.
|
72 |
- `--prompt "..."`: This parameter is used to provide the initial text that you want the model to respond to or complete.
|
73 |
- `--max-tokens`: Sets the maximum number of tokens to generate. For example, `--max-tokens 200` will limit the output to 200 tokens.
|
74 |
- `--temp`: Controls the randomness of the output. Higher temperature values (like 0.8) will produce more diverse and creative outputs, while lower values (like 0.2) will make the output more deterministic and focused. The default value is usually 0.6.
|
|
|
78 |
|
79 |
The following command will use a higher temperature value and limit the output length:
|
80 |
```bash
|
81 |
+
mlx_lm.generate --model openbmb/MiniCPM-8B-mlx \
|
82 |
--prompt "tell me a story about a robot who discovered music" \
|
83 |
--max-tokens 500 \
|
84 |
--temp 0.8
|