Update src/about.py
Browse files- src/about.py +6 -6
src/about.py
CHANGED
@@ -26,10 +26,10 @@ TITLE = """<h1 align="center" id="space-title">π₯ <a href="" target="_blank">U
|
|
26 |
# Links and conference info
|
27 |
LINKS_AND_INFO = """
|
28 |
<div align="center">
|
29 |
-
<p><a href="https://github.com/CodeGoat24/UnifiedReward" target="_blank">UnifiedReward Team</a
|
30 |
|
31 |
-
<a href="
|
32 |
-
<a href="" target="_blank">π arXiv Paper</a>
|
33 |
|
34 |
<a href="https://github.com/CodeGoat24/UniGenBench" target="_blank" rel="noopener noreferrer"><img alt="Code" src="https://img.shields.io/github/stars/CodeGoat24/UniGenBench.svg?style=social&label=Official"></a>
|
35 |
|
@@ -38,9 +38,9 @@ LINKS_AND_INFO = """
|
|
38 |
|
39 |
# What does your leaderboard evaluate?
|
40 |
INTRODUCTION_TEXT = """
|
41 |
-
π [UniGenBench]() is a unified benchmark for T2I generation that integrates diverse prompt themes with a comprehensive suite of fine-grained evaluation criteria.
|
42 |
|
43 |
-
π§ You can use the official [GitHub repo](https://github.com/CodeGoat24/UniGenBench) to evaluate your model on [UniGenBench]().
|
44 |
|
45 |
π To add your own model to the leaderboard, please send an Email to [email protected], then we will help with the evaluation and updating the leaderboard.
|
46 |
"""
|
@@ -51,4 +51,4 @@ INTRODUCTION_TEXT = """
|
|
51 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
52 |
CITATION_BUTTON_TEXT = r"""
|
53 |
|
54 |
-
|
|
|
26 |
# Links and conference info
|
27 |
LINKS_AND_INFO = """
|
28 |
<div align="center">
|
29 |
+
<p><a href="https://github.com/CodeGoat24/UnifiedReward" target="_blank">UnifiedReward</a> Team&<a href="https://hunyuan.tencent.com/" target="_blank">Hunyuan</a>, Tencent</p> <br>
|
30 |
|
31 |
+
<a href="" target="_blank">π Homepage</a> |
|
32 |
+
<a href="" target="_blank">π arXiv Paper</a> | Coming soon!!
|
33 |
|
34 |
<a href="https://github.com/CodeGoat24/UniGenBench" target="_blank" rel="noopener noreferrer"><img alt="Code" src="https://img.shields.io/github/stars/CodeGoat24/UniGenBench.svg?style=social&label=Official"></a>
|
35 |
|
|
|
38 |
|
39 |
# What does your leaderboard evaluate?
|
40 |
INTRODUCTION_TEXT = """
|
41 |
+
π [UniGenBench](https://github.com/CodeGoat24/UniGenBench) is a unified benchmark for T2I generation that integrates diverse prompt themes with a comprehensive suite of fine-grained evaluation criteria.
|
42 |
|
43 |
+
π§ You can use the official [GitHub repo](https://github.com/CodeGoat24/UniGenBench) to evaluate your model on [UniGenBench](https://github.com/CodeGoat24/UniGenBench).
|
44 |
|
45 |
π To add your own model to the leaderboard, please send an Email to [email protected], then we will help with the evaluation and updating the leaderboard.
|
46 |
"""
|
|
|
51 |
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
|
52 |
CITATION_BUTTON_TEXT = r"""
|
53 |
|
54 |
+
"""
|