Update README.md
Browse files
README.md
CHANGED
@@ -1,26 +1,49 @@
|
|
1 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
|
3 |
-
|
4 |
-
|
5 |
-
### Key Features:
|
6 |
|
7 |
-
|
8 |
-
Test LLM applications rigorously across multiple dimensions, including security, bias, reliability, and compliance. Built on industry standards from NIST, MITRE, and OWASP, ensuring robust and defensible evaluations.
|
9 |
|
10 |
-
|
11 |
-
Automatically generate multi-turn, scenario-driven test cases tailored to your application. Test suites dynamically refine based on real-world usage and expert feedback to improve accuracy and relevance.
|
12 |
|
13 |
-
3. **Domain-specific coverage**
|
14 |
-
Leverage pre-built, domain-specific test benches designed to detect sector-specific vulnerabilities in financial services, insurance and more—ensuring reliability and reducing operational risk.
|
15 |
|
16 |
-
|
17 |
-
Stay ahead of emerging threats with automated test updates. Our SDK helps to continuously integrate new adversarial patterns and business-relevant risks, keeping your evaluation process current and effective.
|
18 |
|
19 |
-
|
20 |
-
Run iterative, large-scale test evaluations with minimal setup. Our SDK integrates into CI/CD pipelines, enabling automated, repeatable testing for robust AI validation at scale.
|
21 |
|
22 |
-
|
23 |
-
|
|
|
|
|
24 |
|
25 |
### Example Use Cases:
|
26 |
|
@@ -36,12 +59,6 @@ Enhance collaboration between developers, domain experts, and compliance teams.
|
|
36 |
- **AI Support Chatbot**:
|
37 |
Ensure that your chatbot consistently delivers helpful, accurate, and empathetic responses across various scenarios.
|
38 |
|
39 |
-
### How to Use Our Datasets
|
40 |
-
|
41 |
-
Rhesis AI provides a [SDK on Github](https://github.com/rhesis-ai/rhesis-sdk) and curated selection of datasets for testing LLM applications. These datasets are designed to evaluate the behavior of different types of LLM applications under various conditions. To get started, explore our datasets on Hugging Face, select the relevant test set for your needs, and begin evaluating your applications.
|
42 |
-
|
43 |
-
For more information on how to integrate Rhesis AI into your LLM application testing process, or to inquire about custom test sets, feel free to explore our [Rhesis SDK on Github](https://github.com/rhesis-ai/rhesis-sdk) or reach out to us at: [email protected].
|
44 |
-
|
45 |
### Disclaimer
|
46 |
|
47 |
Our test sets are designed to rigorously evaluate LLM applications across various dimensions, including bias, safety, and security. Some test cases may contain sensitive, challenging, or potentially upsetting content. These cases are included to ensure thorough and realistic assessments. Users should review test cases carefully and exercise discretion when utilizing them.
|
|
|
1 |
+
<p align="center">
|
2 |
+
<img src="https://cdn.prod.website-files.com/66f422128b6d0f3351ce41e3/66fd07dc0b6994070ec5b54b_Logo%20Rhesis%20Orange-p-500.png" alt="Rhesis Logo" width="300"/>
|
3 |
+
</p>
|
4 |
+
|
5 |
+
|
6 |
+
<p align="center">
|
7 |
+
<a href="https://pypi.org/project/rhesis-sdk/">
|
8 |
+
<img src="https://img.shields.io/pypi/v/rhesis-sdk" alt="PyPI Version" style="display:inline-block;">
|
9 |
+
</a>
|
10 |
+
<a href="https://pypi.org/project/rhesis-sdk/">
|
11 |
+
<img src="https://img.shields.io/pypi/pyversions/rhesis-sdk" alt="Python Versions" style="display:inline-block;">
|
12 |
+
</a>
|
13 |
+
<a href="https://discord.rhesis.ai">
|
14 |
+
<img src="https://img.shields.io/discord/1340989671601209408?color=7289da&label=Discord&logo=discord&logoColor=white" alt="Discord" style="display:inline-block;">
|
15 |
+
</a>
|
16 |
+
<a href="https://www.linkedin.com/company/rhesis-ai">
|
17 |
+
<img src="https://img.shields.io/badge/LinkedIn-Rhesis_AI-blue?logo=linkedin" alt="LinkedIn" style="display:inline-block;">
|
18 |
+
</a>
|
19 |
+
<a href="https://huggingface.co/rhesis">
|
20 |
+
<img src="https://img.shields.io/badge/🤗-Rhesis-yellow" alt="Hugging Face" style="display:inline-block;">
|
21 |
+
</a>
|
22 |
+
<a href="https://docs.rhesis.ai">
|
23 |
+
<img src="https://img.shields.io/badge/docs-rhesis.ai-blue" alt="Documentation" style="display:inline-block;">
|
24 |
+
</a>
|
25 |
+
</p>
|
26 |
+
|
27 |
+
|
28 |
+
> Open-source test generation SDK for LLM applications.
|
29 |
+
|
30 |
+
Rhesis enables AI developers to access curated test sets and generate dynamic ones for LLM applications. It provides tools to tailor validations to your needs and integrate seamlessly to keep your Gen AI robust, reliable & compliant.
|
31 |
|
32 |
+
### How to Use Our Datasets
|
|
|
|
|
33 |
|
34 |
+
Rhesis AI provides a [SDK on Github](https://github.com/rhesis-ai/rhesis-sdk) and curated selection of datasets for testing LLM applications. These datasets are designed to evaluate the behavior of different types of LLM applications under various conditions. To get started, explore our datasets on Hugging Face, select the relevant test set for your needs, and begin evaluating your applications.
|
|
|
35 |
|
36 |
+
For more information on how to integrate Rhesis AI into your LLM application testing process, or to inquire about custom test sets, feel free to explore our [Rhesis SDK on Github](https://github.com/rhesis-ai/rhesis-sdk) or reach out to us at: [email protected].
|
|
|
37 |
|
|
|
|
|
38 |
|
39 |
+
## Features
|
|
|
40 |
|
41 |
+
The Rhesis SDK currently provides functionality to work with Rhesis test sets through routine operations:
|
|
|
42 |
|
43 |
+
- **List Test Sets**: Browse through available curated test sets
|
44 |
+
- **Load Test Sets**: Load specific test sets for your use case
|
45 |
+
- **Download Test Sets**: Download test set data for offline use
|
46 |
+
- **Generate Test Sets**: Generate new test sets from basic prompts
|
47 |
|
48 |
### Example Use Cases:
|
49 |
|
|
|
59 |
- **AI Support Chatbot**:
|
60 |
Ensure that your chatbot consistently delivers helpful, accurate, and empathetic responses across various scenarios.
|
61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
### Disclaimer
|
63 |
|
64 |
Our test sets are designed to rigorously evaluate LLM applications across various dimensions, including bias, safety, and security. Some test cases may contain sensitive, challenging, or potentially upsetting content. These cases are included to ensure thorough and realistic assessments. Users should review test cases carefully and exercise discretion when utilizing them.
|