diff --git a/spaces/101-5/gpt4free/g4f/.v1/gui/pywebio-gui/README.md b/spaces/101-5/gpt4free/g4f/.v1/gui/pywebio-gui/README.md
deleted file mode 100644
index 2b99c075d507dbf128a170d2975b1b22b393a70e..0000000000000000000000000000000000000000
--- a/spaces/101-5/gpt4free/g4f/.v1/gui/pywebio-gui/README.md
+++ /dev/null
@@ -1,24 +0,0 @@
-# GUI with PyWebIO
-Simple, fast, and with fewer errors
-Only requires
-```bash
-pip install gpt4free
-pip install pywebio
-```
-clicking on 'pywebio-usesless.py' will run it
-
-PS: Currently, only 'usesless' is implemented, and the GUI is expected to be updated infrequently, with a focus on stability.
-
-↓ Here is the introduction in zh-Hans-CN below.
-
-# 使用pywebio实现的极简GUI
-简单,快捷,报错少
-只需要
-```bash
-pip install gpt4free
-pip install pywebio
-```
-
-双击pywebio-usesless.py即可运行
-
-ps:目前仅实现usesless,这个gui更新频率应该会比较少,目的是追求稳定
diff --git a/spaces/1acneusushi/gradio-2dmoleculeeditor/SEUSv10RC6shaderpackzip ~REPACK~.md b/spaces/1acneusushi/gradio-2dmoleculeeditor/SEUSv10RC6shaderpackzip ~REPACK~.md
deleted file mode 100644
index b1918126a7aa4516d3b4197734d420b7e2442edf..0000000000000000000000000000000000000000
--- a/spaces/1acneusushi/gradio-2dmoleculeeditor/SEUSv10RC6shaderpackzip ~REPACK~.md
+++ /dev/null
@@ -1,126 +0,0 @@
-## SEUSv10RC6shaderpackzip
-
-
-
-
-
-
-
-
-
-**LINK ✸ [https://jinyurl.com/2tzZXd](https://jinyurl.com/2tzZXd)**
-
-
-
-
-
-
-
-
-
-
-
-
-
-# How to Install SEUS v10 RC6 Shader Pack for Minecraft
-
-
-
-SEUS v10 RC6 is a shader pack for Minecraft that enhances the graphics of the game with realistic lighting, shadows, water effects and more. It is one of the most popular shader packs for Minecraft and requires OptiFine or GLSL Shaders Mod to work. In this article, we will show you how to download and install SEUS v10 RC6 shader pack for Minecraft.
-
-
-
-## Step 1: Download SEUS v10 RC6 shader pack
-
-
-
-The first step is to download the SEUS v10 RC6 shader pack from the official website of Sonic Ether[^1^] or from other sources[^2^]. The file name should be `SEUS-v10rc6.zip` and it should be around 20 MB in size. You can also download other versions of SEUS shader pack from the same website, such as SEUS Renewed or SEUS PTGI.
-
-
-
-## Step 2: Install OptiFine or GLSL Shaders Mod
-
-
-
-The next step is to install OptiFine or GLSL Shaders Mod for your Minecraft version. These mods are necessary to run shader packs in Minecraft. You can download OptiFine from its official website or from other sources. You can download GLSL Shaders Mod from its official forum thread or from other sources. Follow the instructions on how to install these mods for your Minecraft version.
-
-
-
-## Step 3: Move the shader pack file to the shader folder
-
-
-
-The final step is to move the `SEUS-v10rc6.zip` file to the shader folder in your Minecraft directory. To do this, open the Minecraft launcher and select the profile that has OptiFine or GLSL Shaders Mod installed. Launch Minecraft and go to the video settings in the options menu. Click on shaders packs and open the shader folder in the lower left corner. This will open a folder called `.minecraft\shaderpacks`. Move the `SEUS-v10rc6.zip` file into this folder. Return to Minecraft and select SEUS v10 RC6 in the shader list. Click on done and enjoy your new graphics!
-
-
-
-## Troubleshooting
-
-
-
-If you encounter any problems with installing or running SEUS v10 RC6 shader pack, here are some possible solutions:
-
-
-
-- Make sure you have a compatible version of OptiFine or GLSL Shaders Mod installed for your Minecraft version.
-
-- Make sure you have a powerful enough computer to run SEUS v10 RC6 shader pack. It is recommended to have at least 4 GB of RAM and a decent graphics card.
-
-- Make sure you have allocated enough memory to Minecraft. You can do this by editing the JVM arguments in the launcher profile settings.
-
-- Make sure you have updated your graphics drivers to the latest version.
-
-- If you still have issues, you can try other versions of SEUS shader pack or other shader packs for Minecraft.
-
-
-
-We hope this article helped you install SEUS v10 RC6 shader pack for Minecraft. If you have any questions or feedback, feel free to leave a comment below.
-
-
-
-## What are the features of SEUS v10 RC6 shader pack?
-
-
-
-SEUS v10 RC6 shader pack is a legacy version of SEUS that offers some amazing features for Minecraft graphics. Some of the features are:
-
-
-
-- Dynamic shadows that change according to the position of the sun and the light sources.
-
-- Realistic water effects that reflect and refract the environment.
-
-- Smooth lighting that eliminates harsh edges and creates soft transitions.
-
-- Bloom and lens flare effects that add a cinematic touch to the scenes.
-
-- Motion blur and depth of field effects that enhance the sense of movement and distance.
-
-- Customizable settings that allow you to adjust the performance and quality of the shader pack.
-
-
-
-SEUS v10 RC6 shader pack is compatible with Minecraft 1.4.6 and requires OptiFine or GLSL Shaders Mod to work. It is one of the most popular shader packs for Minecraft and has been praised by many players and reviewers for its stunning visuals and performance[^1^] [^2^] [^3^].
-
-
-
-## Why should you use SEUS v10 RC6 shader pack?
-
-
-
-If you are looking for a way to improve your Minecraft experience with realistic and immersive graphics, SEUS v10 RC6 shader pack is a great choice. It will transform your Minecraft world into a beautiful and lively place that you can explore and enjoy. You will be amazed by the difference that SEUS v10 RC6 shader pack makes in your game. It will make your Minecraft look like a whole new game.
-
-
-
-SEUS v10 RC6 shader pack is also easy to install and use. You just need to follow the steps in this article and you will be ready to go. You can also customize the settings of the shader pack to suit your preferences and needs. You can change the brightness, contrast, color, fog, water, shadows, motion blur, depth of field and more. You can also toggle some features on and off if you want to save some resources or change the mood of your game.
-
-
-
-SEUS v10 RC6 shader pack is a must-have for any Minecraft fan who wants to experience the game in a new way. It will make your Minecraft more realistic, beautiful and fun. You will not regret trying it out.
-
- 145887f19f
-
-
-
-
-
diff --git a/spaces/1gistliPinn/ChatGPT4/Examples/Articulate Storyline 212121412 Portable.md b/spaces/1gistliPinn/ChatGPT4/Examples/Articulate Storyline 212121412 Portable.md
deleted file mode 100644
index 12a338023374674e1ded0fd398767709ea7b4b11..0000000000000000000000000000000000000000
--- a/spaces/1gistliPinn/ChatGPT4/Examples/Articulate Storyline 212121412 Portable.md
+++ /dev/null
@@ -1,22 +0,0 @@
-
-
-The flexible storyline is fully integrated into the POWER BI interface and provides an intuitive user interface for users to start, pause, review and resume...
-
-A well written and illustrated biography of John Calhoun, first Vice President of the United States of America. He was also the first Vice President to become President of the United States after his brother, Andrew Johnson died. Calhoun was also a great Civil War general. Learn about the best of America in this...
-
-It's always a struggle to be "the" girl and a budding athlete in the same group of girls, but can all three teenage girls really have it all? This is the story of the relationships these three girls form throughout their senior year of high school and how their friendship not only helps them get through the...
-
-How the media is misleading US citizens on the separation of powers. "Meet my Friend, Nancy Pelosi." She says the media is wrong. "Our representatives are working on our behalf." It's not the same way, Nancy Pelosi says. The media is twisting the facts. She knows. So is Harry Reid. The media is just...
-
-One of the most groundbreaking books in American history!The story of the evolution of democracy from its beginning, through our founding fathers. This is a book that has history shaped around you as you read it. It gives a powerful overview of the history of our country. You will meet President...
-
-Warren is a small town, a peaceful town. But every summer when the town seems to sleep, there are mysterious happenings. With a new library and community center, things have never been better. But when the new summer arrives and the town seems to sleep, something wakes up in Warren and people start...
-
-It all began on May 20th 2008, what started as a simple wager on an online gaming site. He lost and was then left for dead. With nothing and nowhere to go he landed in a nother world; the world of supernatural witches. It was there he found his true home and where he would begin his journey as...
-
-It was only a little over a year ago when Stacey Little left her small, quiet, Canadian town. She left with a memory that would haunt her forever. Not long after that, she started to hear voices and see visions. She didn't know what to make of it. She tried to ignore it, but the voices and...
-
-"Just following the light" In the year 2000, American's were a bit scared 4fefd39f24
-
-
-
diff --git a/spaces/1gistliPinn/ChatGPT4/Examples/Ek Bura Aadmi English Dubbed Download A Review of the Film Based on a True Story.md b/spaces/1gistliPinn/ChatGPT4/Examples/Ek Bura Aadmi English Dubbed Download A Review of the Film Based on a True Story.md
deleted file mode 100644
index 4fc77f22ced9978d8f98e302abebfb37166b247f..0000000000000000000000000000000000000000
--- a/spaces/1gistliPinn/ChatGPT4/Examples/Ek Bura Aadmi English Dubbed Download A Review of the Film Based on a True Story.md
+++ /dev/null
@@ -1,6 +0,0 @@
-
-
- aaccfb2cb3
-
-
-
diff --git a/spaces/1line/AutoGPT/tests/integration/milvus_memory_tests.py b/spaces/1line/AutoGPT/tests/integration/milvus_memory_tests.py
deleted file mode 100644
index ec38bf2f72087b5da679d26594ebff97d8a09b19..0000000000000000000000000000000000000000
--- a/spaces/1line/AutoGPT/tests/integration/milvus_memory_tests.py
+++ /dev/null
@@ -1,57 +0,0 @@
-# sourcery skip: snake-case-functions
-"""Tests for the MilvusMemory class."""
-import random
-import string
-import unittest
-
-from autogpt.config import Config
-from autogpt.memory.milvus import MilvusMemory
-
-try:
-
- class TestMilvusMemory(unittest.TestCase):
- """Tests for the MilvusMemory class."""
-
- def random_string(self, length: int) -> str:
- """Generate a random string of the given length."""
- return "".join(random.choice(string.ascii_letters) for _ in range(length))
-
- def setUp(self) -> None:
- """Set up the test environment."""
- cfg = Config()
- cfg.milvus_addr = "localhost:19530"
- self.memory = MilvusMemory(cfg)
- self.memory.clear()
-
- # Add example texts to the cache
- self.example_texts = [
- "The quick brown fox jumps over the lazy dog",
- "I love machine learning and natural language processing",
- "The cake is a lie, but the pie is always true",
- "ChatGPT is an advanced AI model for conversation",
- ]
-
- for text in self.example_texts:
- self.memory.add(text)
-
- # Add some random strings to test noise
- for _ in range(5):
- self.memory.add(self.random_string(10))
-
- def test_get_relevant(self) -> None:
- """Test getting relevant texts from the cache."""
- query = "I'm interested in artificial intelligence and NLP"
- num_relevant = 3
- relevant_texts = self.memory.get_relevant(query, num_relevant)
-
- print(f"Top {k} relevant texts for the query '{query}':")
- for i, text in enumerate(relevant_texts, start=1):
- print(f"{i}. {text}")
-
- self.assertEqual(len(relevant_texts), k)
- self.assertIn(self.example_texts[1], relevant_texts)
-
-except:
- print(
- "Skipping tests/integration/milvus_memory_tests.py as Milvus is not installed."
- )
diff --git a/spaces/1phancelerku/anime-remove-background/Discover the New Champion Stadium in Pokmon Masters EX on GBA.md b/spaces/1phancelerku/anime-remove-background/Discover the New Champion Stadium in Pokmon Masters EX on GBA.md
deleted file mode 100644
index 1561bfb224e719367f600e116b7baf5687ffa448..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Discover the New Champion Stadium in Pokmon Masters EX on GBA.md
+++ /dev/null
@@ -1,147 +0,0 @@
-
-
Pokemon Masters EX: A New Adventure with Your Favorite Trainers and Pokemon
-
If you are a fan of Pokemon games, you might have heard of Pokemon Masters EX, a free-to-play mobile game that features an original story, team-ups between iconic trainers and Pokemon, and exciting 3-on-3 battles. Whether you are new to the game or a veteran player, you might be wondering how to download and play Pokemon Masters EX on your device. In this article, we will show you how to do that, as well as give you some tips and tricks for enjoying the game.
-
Pokemon Masters EX is a spin-off game that takes place on the artificial island of Pasio, where trainers from different regions compete in the Pokemon Masters League (PML), a tournament of 3-on-3 battles. You play as a trainer who teams up with other famous trainers and their Pokemon, called sync pairs, to form a team of three. You can also participate in co-op battles with other players, as well as special events that feature characters from the anime series.
Some of the main features of Pokemon Masters EX are:
-
-
Sync pairs: You can collect over 200 sync pairs, each with their own unique moves, skills, and personalities. You can also upgrade their potential, unlock their sync grids, and change their outfits.
-
3-on-3 battles: You can experience a new kind of Pokemon battle that is fast-paced, strategic, and fun. You can use moves, items, sync moves, and passive skills to defeat your opponents.
-
Champion Stadium: You can challenge the Elite Four and regional champions of different regions in a weekly challenge. You can choose the order and difficulty of the battles and earn rewards based on your performance.
-
Events: You can join various events that feature story episodes, missions, rewards, and exclusive sync pairs. Some events are based on the seasons, holidays, or special occasions.
-
-
Now that you know what Pokemon Masters EX is and why you should play it, let's see how you can download and play it on your device.
-
How to Download Pokemon Masters EX on Android
-
If you have an Android device, you can easily download and play Pokemon Masters EX from the Google Play Store. Here are the steps you need to follow:
-
pokemon masters ex gba rom download
-pokemon masters ex gba emulator download
-pokemon masters ex gba hack download
-pokemon masters ex gba apk download
-pokemon masters ex gba cheats download
-pokemon masters ex gba mod download
-pokemon masters ex gba free download
-pokemon masters ex gba android download
-pokemon masters ex gba ios download
-pokemon masters ex gba pc download
-how to download pokemon masters ex gba
-where to download pokemon masters ex gba
-best site to download pokemon masters ex gba
-safe download pokemon masters ex gba
-fast download pokemon masters ex gba
-easy download pokemon masters ex gba
-full version download pokemon masters ex gba
-latest version download pokemon masters ex gba
-updated version download pokemon masters ex gba
-offline download pokemon masters ex gba
-online play pokemon masters ex gba without download
-review of pokemon masters ex gba download
-guide for pokemon masters ex gba download
-tips for pokemon masters ex gba download
-tricks for pokemon masters ex gba download
-walkthrough for pokemon masters ex gba download
-gameplay of pokemon masters ex gba download
-features of pokemon masters ex gba download
-characters of pokemon masters ex gba download
-story of pokemon masters ex gba download
-graphics of pokemon masters ex gba download
-sound of pokemon masters ex gba download
-controls of pokemon masters ex gba download
-compatibility of pokemon masters ex gba download
-performance of pokemon masters ex gba download
-quality of pokemon masters ex gba download
-rating of pokemon masters ex gba download
-ranking of pokemon masters ex gba download
-popularity of pokemon masters ex gba download
-demand of pokemon masters ex gba download
-benefits of pokemon masters ex gba download
-advantages of pokemon masters ex gba download
-disadvantages of pokemon masters ex gba download
-drawbacks of pokemon masters ex gba download
-problems of pokemon masters ex gba download
-issues of pokemon masters ex gba download
-solutions of pokemon masters ex gba download
-alternatives of pokemon masters ex gba download
-comparisons of pokemon masters ex gba download
-
-
Go to the Google Play Store and search for Pokemon Masters EX. You can also use this link: [Pokemon Masters EX].
-
Tap on the Install button and wait for the game to download. The game size is about 2 GB, so make sure you have enough space and a stable internet connection.
-
Launch the game and follow the instructions to start your adventure. You will need to accept the terms of service, choose your language, download additional data, and create your character.
-
-
Congratulations! You are now ready to play Pokemon Masters EX on your Android device.
-
How to Download Pokemon Masters EX on iOS
-
If you have an iOS device, you can also download and play Pokemon Masters EX from the App Store. Here are the steps you need to follow:
-
-
Go to the App Store and search for Pokemon Masters EX. You can also use this link: [Pokemon Masters EX].
-
Tap on the Get button and enter your Apple ID password if prompted. The game is free to download, but it may offer in-app purchases.
-
Wait for the game to download and launch it from your home screen. You will need to accept the terms of service, choose your language, download additional data, and create your character.
-
-
Congratulations! You are now ready to play Pokemon Masters EX on your iOS device.
How to Download Pokemon Masters EX on PC
-
If you don't have a mobile device or you prefer to play Pokemon Masters EX on a bigger screen, you can also download and play it on your PC using an Android emulator. An Android emulator is a software that allows you to run Android apps and games on your PC. Here are the steps you need to follow:
-
-
Download and install an Android emulator such as [BlueStacks] or [NoxPlayer]. You can choose the one that suits your preferences and system requirements.
-
Open the emulator and sign in with your Google account. You will need to do this to access the Google Play Store and other Google services.
-
Go to the Google Play Store and search for Pokemon Masters EX. You can also use this link: [Pokemon Masters EX].
-
Install the game and launch it from the emulator. You will need to accept the terms of service, choose your language, download additional data, and create your character.
-
-
Congratulations! You are now ready to play Pokemon Masters EX on your PC.
-
How to Download Pokemon Masters EX on GBA
-
If you are feeling nostalgic or adventurous, you might want to try playing Pokemon Masters EX on a Game Boy Advance (GBA), a handheld console that was released in 2001. GBA is one of the most popular and beloved gaming devices of all time, and it has a huge library of classic games, including many Pokemon titles. However, Pokemon Masters EX is not officially available for GBA, so you will need to use some tricks to make it work.
-
Disclaimer: Before we show you how to download and play Pokemon Masters EX on GBA, we need to warn you about the legal and ethical issues of using GBA emulators and ROMs. Emulators are software that mimic the hardware and software of a console, while ROMs are files that contain the data of a game. Using emulators and ROMs without owning the original console and game is considered piracy and may violate the intellectual property rights of the developers and publishers. Therefore, we do not condone or encourage the use of emulators and ROMs for any illegal or unethical purposes. If you decide to use them, do so at your own risk and responsibility.
-
Here are the steps you need to follow:
-
-
Download and install a GBA emulator such as [VisualBoyAdvance] or [mGBA]. These are two of the most popular and reliable GBA emulators for PC.
-
Download a Pokemon Masters EX ROM from a reputable source such as [ROMsMania] or [EmuParadise]. These are two of the most trusted and safe websites for downloading ROMs for various consoles.
-
Open the emulator and load the ROM file. You will need to locate the ROM file on your PC and open it with the emulator.
-
Enjoy playing Pokemon Masters EX on your GBA. You can customize the controls, graphics, sound, and other settings according to your preferences.
-
-
Congratulations! You are now ready to play Pokemon Masters EX on your GBA.
Pros and Cons of Playing Pokemon Masters EX on GBA
-
Playing Pokemon Masters EX on GBA might sound like a fun and nostalgic idea, but it also has its advantages and disadvantages. Here are some of them:
-
-
-
Pros
-
Cons
-
-
-
- You can experience the game in a retro style, with pixelated graphics and chiptune sound.
-
- You will miss out on the high-quality graphics and sound of the original game, which are designed to enhance the gameplay and immersion.
-
-
-
- You can play the game on a portable device, without needing an internet connection or a battery charger.
-
- You will need to carry around a GBA device and a cartridge, which might be inconvenient or impractical in some situations.
-
-
-
- You can customize the game settings, such as the speed, difficulty, cheats, and save states, to suit your preferences and needs.
-
- You might encounter compatibility issues, bugs, glitches, or crashes that could ruin your gaming experience or damage your device.
-
-
-
- You can enjoy the game in a different way, with new challenges and surprises.
-
- You will not be able to access the latest updates, features, events, and sync pairs that are available in the original game.
-
-
-
As you can see, playing Pokemon Masters EX on GBA has its pros and cons, and it is up to you to decide whether it is worth it or not. If you do decide to try it, make sure you do it legally and ethically, and have fun!
-
Tips and Tricks for Playing Pokemon Masters EX
-
Pokemon Masters EX is a game that requires strategy, skill, and knowledge to master. If you want to become a better player and enjoy the game more, here are some tips and tricks that might help you:
-
-
How to build a balanced team of sync pairs: A good team should consist of three sync pairs that complement each other's strengths and weaknesses. You should have one striker (a sync pair that deals high damage), one support (a sync pair that boosts the stats or heals the team), and one tech (a sync pair that debuffs the enemy or inflicts status effects). You should also consider the type, role, and synergy of your sync pairs when choosing them.
-
How to level up and unlock new moves and skills: To level up your sync pairs, you need to use items such as level-up manuals, buff blends, aid ades, and tech tonics. You can obtain these items from various sources, such as battles, missions, events, and shops. To unlock new moves and skills for your sync pairs, you need to use items such as move candies, skill capsules, and sync orbs. You can also obtain these items from various sources, such as battles, missions, events, and shops.
How to use sync moves and sync grids effectively: Sync moves are powerful moves that can be used once the sync move countdown reaches zero. Sync moves can deal massive damage, boost your stats, or trigger other effects depending on your sync pair. Sync grids are skill trees that allow you to customize your sync pairs' abilities and unlock new passive skills, sync move effects, or lucky skills. To use sync moves and sync grids effectively, you need to plan ahead and choose the ones that suit your team and strategy.
-
How to participate in co-op battles and events: Co-op battles are battles that allow you to team up with other players online and fight against a common enemy. Co-op battles can be more challenging and rewarding than solo battles, as they require coordination, communication, and cooperation. Events are limited-time activities that offer exclusive story episodes, missions, rewards, and sync pairs. Events can be solo or co-op, and they often have different difficulties and requirements. To participate in co-op battles and events, you need to have an internet connection and a compatible device.
-
How to earn gems and other resources: Gems are the premium currency of the game, which can be used to scout for new sync pairs or exchange for other items. You can earn gems by completing various tasks, such as logging in daily, completing missions, participating in events, or purchasing them with real money. Other resources include coins, vouchers, tickets, medals, and more. You can earn these resources by playing the game normally, or by exchanging them with gems or other items.
-
-
Conclusion
-
Pokemon Masters EX is a game that offers a lot of fun and excitement for Pokemon fans and gamers alike. You can download and play it on various devices, such as Android, iOS, PC, or even GBA. You can also enjoy the game's features, such as sync pairs, 3-on-3 battles, Champion Stadium, events, and more. However, you also need to be aware of the legal and ethical issues of using emulators and ROMs, as well as the pros and cons of playing the game on different devices. We hope this article has helped you learn how to download and play Pokemon Masters EX on your device of choice. If you have any questions or feedback, feel free to leave a comment below. Thank you for reading and happy gaming!
-
FAQs
-
Here are some frequently asked questions about Pokemon Masters EX:
-
-
Q: Is Pokemon Masters EX free to play?
-A: Yes, Pokemon Masters EX is free to download and play. However, it may offer in-app purchases that can enhance your gaming experience.
-
Q: Is Pokemon Masters EX compatible with my device?
-A: Pokemon Masters EX is compatible with most Android devices that have Android OS 7.0 or higher (64-bit), and most iOS devices that have iOS 11 or higher. For PC and GBA devices, you will need to use an emulator and a ROM file.
-
Q: How can I get more sync pairs in Pokemon Masters EX?
-A: You can get more sync pairs by scouting them using gems or tickets. You can also get some sync pairs by completing story chapters or events.
-
Q: How can I contact the support team of Pokemon Masters EX?
-A: You can contact the support team of Pokemon Masters EX by tapping on the Menu button in the game, then tapping on Other > Customer Support > Inquiries.
-
Q: How can I join the community of Pokemon Masters EX?
-A: You can join the community of Pokemon Masters EX by following their official social media accounts, such as [Facebook], [Twitter], [Instagram], [YouTube], or [Reddit]. You can also join their official [Discord] server or their [website].
- 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Download Classic Solitaire for Windows - Free and Easy.md b/spaces/1phancelerku/anime-remove-background/Download Classic Solitaire for Windows - Free and Easy.md
deleted file mode 100644
index 2bcceb4830e7f1227fae226350ffbfe65f21a04f..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Download Classic Solitaire for Windows - Free and Easy.md
+++ /dev/null
@@ -1,116 +0,0 @@
-
-
Classic Solitaire Download for PC: How to Play the Timeless Card Game on Your Windows Device
-
If you are looking for a simple yet addictive game that you can play on your PC, you might want to try classic solitaire. This card game has been around for centuries and has entertained millions of people around the world. In this article, we will show you how to download and install classic solitaire for PC, what features it offers, and some tips and tricks to help you win more games.
Classic solitaire, also known as Klondike solitaire, is a card game that involves arranging 52 cards into four piles, called foundations, according to their suits and ranks. The game starts with seven columns of cards, called tableau, with the top card of each column face up and the rest face down. The goal is to move all the cards from the tableau and the stock pile (the remaining cards that are not dealt) to the foundations, starting from the ace and ending with the king.
-
Why play classic solitaire on PC?
-
Classic solitaire is a game that can be played by anyone, regardless of age or skill level. It is a great way to pass the time, relax, and exercise your brain. Playing classic solitaire on PC has some advantages over playing it on other devices, such as:
-
-
You can enjoy a larger screen and better graphics.
-
You can use your mouse or keyboard to control the game.
-
You can access more features and options, such as different game modes, themes, statistics, and achievements.
-
You can play online or offline, without worrying about battery life or internet connection.
-
-
How to download and install classic solitaire for PC?
-
There are many ways to download and install classic solitaire for PC, but one of the easiest and most reliable methods is to use the Microsoft Store. Here are the steps to follow:
-
free classic solitaire game download for windows 10
-classic solitaire card game download for pc offline
-how to download classic solitaire on windows 7
-classic spider solitaire download for pc full version
-classic solitaire no ads download for pc
-classic klondike solitaire download for windows 8
-microsoft classic solitaire collection download for pc
-classic solitaire for pc free download without internet
-best classic solitaire app download for windows 10
-classic solitaire download for pc windows xp
-old classic solitaire game download for pc
-classic pyramid solitaire download for windows 10
-classic solitaire plus download for pc
-where to download classic solitaire for windows 7
-classic tripeaks solitaire download for pc
-original classic solitaire download for windows 10
-classic freecell solitaire download for pc
-classic mahjong solitaire download for windows 10
-classic solitaire hd download for pc
-classic golf solitaire download for windows 10
-easy classic solitaire game download for pc
-classic spider solitaire free download for windows 7
-microsoft classic solitaire free download for pc
-classic hearts solitaire download for windows 10
-play classic solitaire online free no download for pc
-classic minesweeper and solitaire download for windows 10
-classic yukon solitaire download for pc
-new classic solitaire game download for windows 10
-classic canfield solitaire download for pc
-install classic solitaire on windows 10 free download
-simple classic solitaire game download for pc
-old school classic solitaire free download for pc
-microsoft store classic solitaire download for windows 10
-fun classic solitaire games free download for pc
-fast and easy classic solitaire free download for pc
-best free classic spider solitaire download for windows 10
-play store classic solitaire game free download for pc
-microsoft original classic solitaire free download for pc
-cool and relaxing classic solitaire game free download for pc
-addictive and challenging classic freecell solitaire free download for pc
-beautiful and smooth classic mahjong solitaire free download for pc
-awesome and exciting classic tripeaks solitaire free download for pc
-enjoy the old fashioned classic klondike solitaire free download for pc
-learn how to play the ancient game of classic pyramid solitaire free download for pc
-test your skills with the tricky and strategic game of classic canfield solitaire free download for pc
-have fun with the popular and colorful game of classic golf solitaire free download for pc
-experience the thrill of the wild and unpredictable game of classic yukon solitaire free download for pc
On the Microsoft Solitaire Collection page in Microsoft Store, select Install. The game will download and install automatically.
-
To launch the game, select Play. You can always launch the game from the product page, but there's an easier way--pin it. With the game open, right-click (or press and hold) the game button on your task bar and select Pin to task bar . When you close the game, the button will still be there.
-
On the Start menu, scroll down the all apps list to Microsoft Solitaire Collection, right-click (or press and hold) the tile and select Pin to Start . It'll be available on the Start menu.
-
-
If you have any problems with downloading or installing the game, you can run the app troubleshooter or contact Microsoft support for help.
-
Features of classic solitaire for PC
-
Different game modes and difficulty levels
-
One of the best features of classic solitaire for PC is that it offers different game modes and difficulty levels to suit your preference and challenge. You can choose from five game modes: Klondike, Spider, FreeCell, Pyramid, and TriPeaks. Each game mode has its own rules and strategies, so you can try them all and find your favorite one. You can also adjust the difficulty level of each game mode, from easy to expert, depending on how confident you are with your solitaire skills. You can change the game mode and difficulty level anytime from the settings menu.
-
Customizable themes and card backs
-
Another feature of classic solitaire for PC is that it allows you to customize the appearance of the game according to your taste. You can choose from different themes and card backs to make the game more colorful and fun. You can select from various themes, such as nature, animals, sports, holidays, and more. You can also pick from different card backs, such as classic, modern, vintage, and more. You can change the theme and card back anytime from the settings menu.
-
Statistics and achievements
-
If you are a competitive solitaire player, you will love the statistics and achievements feature of classic solitaire for PC. This feature lets you track your progress and performance in the game, such as how many games you have played, won, and lost, how long it took you to finish a game, what your best score and streak are, and more. You can also earn achievements by completing certain goals or challenges in the game, such as winning a game without using undo, clearing all the cards in the tableau, or finishing a game in less than a minute. You can view your statistics and achievements anytime from the main menu.
-
Online and offline play
-
One of the most convenient features of classic solitaire for PC is that it supports both online and offline play. This means that you can play the game anytime and anywhere, whether you have an internet connection or not. When you play online, you can access more features and benefits, such as daily challenges, events, leaderboards, cloud saving, and more. When you play offline, you can still enjoy the basic features of the game, such as different game modes, difficulty levels, themes, and card backs. You can switch between online and offline play anytime from the settings menu.
-
Tips and tricks for playing classic solitaire on PC
-
Use the undo button wisely
-
One of the most useful tools in classic solitaire for PC is the undo button. This button allows you to undo your last move or action in case you make a mistake or change your mind. However, you should not rely on this button too much or use it randomly. You should use it strategically and sparingly, as it can affect your score and time. For example, you should use it when you realize that you have missed a better move or when you want to explore a different option.
-
Pay attention to the cards in the stock pile
-
Another tip for playing classic solitaire on PC is to pay attention to the cards in the stock pile. The stock pile is where the remaining cards that are not dealt are placed. You can draw one or three cards from the stock pile at a time, depending on your difficulty level. You should keep an eye on the cards in the stock pile, as they can help you plan your moves ahead and avoid getting stuck. For example, if you know that there is an ace or a two in the stock pile, you can save a space for it in the foundation.
-
Move cards to the foundation as soon as possible
-
Another tip for playing classic solitaire on PC is to move cards to the foundation as soon as possible. The foundation is where you place the cards in ascending order according to their suits and ranks. The sooner you move cards to the foundation, the easier it will be to clear the tableau and win the game. Moving cards to the foundation also frees up space in the tableau and gives you more options for moving other cards around.
-
Try to clear the columns with the most cards first
-
Another tip for playing classic solitaire on PC is to try to clear the columns with the most cards first. The columns are the vertical stacks of cards in the tableau. The more cards you have in a column, the harder it will be to move them around and access the cards underneath. Therefore, you should try to clear the columns with the most cards first, especially if they have face-down cards. This will help you reveal more cards and create more empty spaces in the tableau.
-
Use the hint button if you get stuck
-
Another tip for playing classic solitaire on PC is to use the hint button if you get stuck. The hint button is located at the bottom right corner of the screen and it will show you a possible move that you can make. However, you should not use the hint button too often or blindly follow its suggestions. You should use it only when you have no other moves or when you want to check if you have missed something. Using the hint button too much can lower your score and make the game less fun.
-
Conclusion
-
Summary of the main points
-
In conclusion, classic solitaire is a timeless card game that you can play on your PC for free. It is a simple yet addictive game that can help you pass the time, relax, and exercise your brain. To play classic solitaire on PC, you can download and install it from the Microsoft Store. You can enjoy different game modes, difficulty levels, themes, card backs, statistics, achievements, and online and offline play. You can also improve your solitaire skills by following some tips and tricks, such as using the undo button wisely, paying attention to the cards in the stock pile, moving cards to the foundation as soon as possible, clearing the columns with the most cards first, and using the hint button if you get stuck.
-
Call to action
-
If you are ready to play classic solitaire on PC, what are you waiting for? Download and install it now and start having fun with this classic card game. You can also share your thoughts and experiences with us in the comments section below. We would love to hear from you!
-
Frequently Asked Questions
-
What are the rules of classic solitaire?
-
The rules of classic solitaire are simple: you have to arrange 52 cards into four piles, called foundations, according to their suits and ranks. The game starts with seven columns of cards, called tableau, with the top card of each column face up and the rest face down. The goal is to move all the cards from the tableau and the stock pile (the remaining cards that are not dealt) to the foundations, starting from the ace and ending with the king.
-
How do I win classic solitaire?
-
You win classic solitaire when you move all the cards from the tableau and the stock pile to the foundations. You can move cards from one place to another by following these rules:
-
-
You can only move one card at a time.
-
You can only move a card that is face up.
-
You can only move a card to an empty space or on top of another card that is one rank higher and of a different color.
-
You can only move a card to a foundation if it matches its suit and rank.
-
You can draw one or three cards from the stock pile at a time, depending on your difficulty level.
-
-
How do I change the game mode or difficulty level?
-
You can change the game mode or difficulty level anytime from the settings menu. To access the settings menu, click on the gear icon at the top right corner of the screen. Then, you can select the game mode and difficulty level that you want to play. You can choose from five game modes: Klondike, Spider, FreeCell, Pyramid, and TriPeaks. You can also adjust the difficulty level of each game mode, from easy to expert.
-
How do I customize the theme or card back?
-
You can customize the theme or card back anytime from the settings menu. To access the settings menu, click on the gear icon at the top right corner of the screen. Then, you can select the theme and card back that you want to use. You can choose from different themes and card backs to make the game more colorful and fun. You can select from various themes, such as nature, animals, sports, holidays, and more. You can also pick from different card backs, such as classic, modern, vintage, and more.
-
How do I view my statistics or achievements?
-
You can view your statistics or achievements anytime from the main menu. To access the main menu, click on the hamburger icon at the top left corner of the screen. Then, you can select the statistics or achievements option that you want to see. You can view your progress and performance in the game, such as how many games you have played, won, and lost, how long it took you to finish a game, what your best score and streak are, and more. You can also earn achievements by completing certain goals or challenges in the game, such as winning a game without using undo, clearing all the cards in the tableau, or finishing a game in less than a minute.
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Download Dream League Soccer 2021 Hack APK and Get Unlimited Coins and Diamonds.md b/spaces/1phancelerku/anime-remove-background/Download Dream League Soccer 2021 Hack APK and Get Unlimited Coins and Diamonds.md
deleted file mode 100644
index b180df7ddefafaff29a573dc9ffa09c24d5cb8b7..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Download Dream League Soccer 2021 Hack APK and Get Unlimited Coins and Diamonds.md
+++ /dev/null
@@ -1,93 +0,0 @@
-
-
How to Hack Dream League Soccer 2021 and Get Unlimited Coins
-
Dream League Soccer 2021 is one of the most popular soccer games on Android and iOS devices. It lets you create your own team, compete in various leagues, and customize your players, kits, stadiums, and more. However, the game also has some limitations, such as limited coins, ads, and in-app purchases that can affect your gaming experience.
-
dream league soccer 2021 hack monedas infinitas apk 2021
If you want to enjoy Dream League Soccer 2021 without any restrictions, you might be interested in hacking the game and getting unlimited coins. With this hack, you can buy any player you want, upgrade your facilities, and unlock all the features of the game. Sounds amazing, right?
-
In this article, we will show you how to download and install the Dream League Soccer 2021 hack apk file, how to use the hack features, and what are the benefits and risks of using it. We will also give you some tips and tricks to play the game like a pro. So, let's get started!
-
How to Download and Install the Dream League Soccer 2021 Hack Apk File
-
The first step to hack Dream League Soccer 2021 is to download and install the hack apk file. This is a modified version of the original game that has been tweaked to give you unlimited coins and other advantages. Here are the steps to follow:
-
-
Go to a reliable website that offers the Dream League Soccer 2021 hack apk file. For example, you can visit [Tablet Adam](^1^), which is a trusted source for Android games and apps.
-
Click on the download button and wait for the file to be downloaded on your device.
-
Before installing the file, make sure you have enabled the "Unknown Sources" option in your device settings. This will allow you to install apps from sources other than the Google Play Store.
-
Locate the downloaded file in your file manager and tap on it to start the installation process.
-
Follow the instructions on the screen and wait for the installation to be completed.
-
Launch the game and enjoy!
-
-
How to Use the Hack Features and Enjoy the Game
-
Now that you have installed the Dream League Soccer 2021 hack apk file, you can use its features and enjoy the game. Here are some of the things you can do with the hack:
-
dream league soccer 2021 mod apk unlimited coins and gems
-descargar dream league soccer 2021 hackeado monedas infinitas
-dream league soccer 2021 cheats android no root
-como hackear dream league soccer 2021 sin root
-dream league soccer 2021 hack ios download
-dream league soccer 2021 unlimited money and players
-baixar dream league soccer 2021 hack dinheiro infinito
-dream league soccer 2021 hack online generator
-dream league soccer 2021 hack apk obb data
-dream league soccer 2021 mod menu apk download
-dream league soccer 2021 hack version free download
-como tener monedas infinitas en dream league soccer 2021 sin aplicaciones
-dream league soccer 2021 hack apk mediafıre
-dream league soccer 2021 mod apk all players unlocked
-dream league soccer 2021 cheat codes for android
-comment hacker dream league soccer 2021 sans verification humaine
-dream league soccer 2021 hack apk latest version
-dream league soccer 2021 mod apk unlimited everything
-como conseguir monedas infinitas en dream league soccer 2021 facil y rapido
-dream league soccer 2021 hack tool no survey no password
-download game dream league soccer 2021 mod apk unlimited money
-trucos para dream league soccer 2021 monedas infinitas
-cara hack dream league soccer 2021 tanpa root
-dream league soccer 2021 hack apk android oyun club
-how to hack dream league soccer 2021 with lucky patcher
-telecharger dream league soccer 2021 mod apk argent illimité
-como instalar dream league soccer 2021 hackeado monedas infinitas
-dream league soccer 2021 mod apk revdl
-how to get unlimited coins in dream league soccer 2021 without human verification
-descargar e instalar dream league soccer 2021 hack monedas infinitas apk gratis
-how to hack players in dream league soccer 2021 ios
-como baixar e instalar dream league soccer 2021 com dinheiro infinito
-download game mod apk offline dream league soccer 2021 unlimited coins and gems
-como tener jugadores al maximo en dream league soccer 2021 sin hackearlo
-how to download and install dream league soccer 2021 mod apk unlimited money and players unlocked
-descargar e instalar el juego de futbol mas popular del mundo: Dream League Soccer 2021 Hack Monedas Infinitas Apk Gratis Para Android y iOS.
-
-
You will have unlimited coins in your account. You can use them to buy any player you want, upgrade your facilities, customize your kits, stadiums, logos, etc.
-
You will not see any ads or in-app purchases in the game. You can play without any interruptions or distractions.
-
You will not need to root or jailbreak your device to run the hack. It works on any Android or iOS device without any problems.
-
-
To use these features, you just need to play the game as usual. You will see that everything is unlocked and available for you. You can also access the settings menu and adjust some options
Benefits of Using Dream League Soccer 2021 Hack Apk
-
Using the Dream League Soccer 2021 hack apk can give you a lot of benefits that can make your gaming experience more fun and satisfying. Here are some of the benefits you can enjoy with the hack:
-
-
Unlimited coins: Coins are the main currency in Dream League Soccer 2021, and you need them to buy players, kits, stadiums, and other items. With the hack, you will have unlimited coins in your account, which means you can buy any player you want, from Lionel Messi to Cristiano Ronaldo, and create your dream team. You can also upgrade your facilities, such as your stadium, medical center, training ground, and academy, to improve your team's performance and reputation. You can also customize your kits, logos, and banners to suit your style and preferences.
-
No ads or in-app purchases: Dream League Soccer 2021 is a free-to-play game, but it also has ads and in-app purchases that can interrupt your gameplay or tempt you to spend real money. With the hack, you will not see any ads or in-app purchases in the game, which means you can play without any distractions or pressure. You can also save your money and use it for other purposes.
-
No root or jailbreak required: Some hacks require you to root or jailbreak your device to run them, which can be risky and complicated. However, the Dream League Soccer 2021 hack apk does not require any root or jailbreak, which means you can run it on any Android or iOS device without any problems. You just need to download and install the file, and you're good to go.
-
-
Risks of Using Dream League Soccer 2021 Hack Apk
-
While using the Dream League Soccer 2021 hack apk can have many benefits, it also has some risks that you should be aware of. Here are some of the risks you may face with the hack:
-
-
Possible ban from the game server or Google Play Store: Using a hack apk file is against the terms and conditions of Dream League Soccer 2021 and Google Play Store. If the game developers or Google detect that you are using a hack, they may ban your account from the game server or Google Play Store. This means you will not be able to play the game or access other apps and services from Google. You may also lose your game progress and data.
-
Possible malware or virus infection from unknown sources: Downloading and installing a hack apk file from unknown sources can be dangerous, as they may contain malware or viruses that can harm your device or steal your personal information. You should always be careful when downloading and installing files from unknown sources, and use a reliable antivirus software to scan them before opening them.
-
Possible loss of game progress or data corruption: Using a hack apk file may also affect your game progress or data, as it may not be compatible with the latest version of the game or your device. You may experience crashes, glitches, errors, or data corruption that can ruin your gaming experience or cause you to lose your game progress and data.
-
Tips and Tricks to Play Dream League Soccer 2021 Like a Pro
-
If you want to play Dream League Soccer 2021 like a pro, you don't need to rely on hacks or cheats. You can improve your skills and tactics by following some tips and tricks that can help you win more matches and trophies. Here are some of them:
-
-
How to build your dream team and manage your squad: The first thing you need to do is to build your dream team and manage your squad. You can choose from over 4,000 licensed players from different leagues and countries, and create your own formations, tactics, and styles. You can also scout new players, transfer players, renew contracts, and train your players to improve their attributes and skills. You should always keep an eye on your players' morale, fitness, and chemistry, as they can affect their performance on the pitch.
-
How to improve your skills and tactics in different modes: The next thing you need to do is to improve your skills and tactics in different modes. You can play in various modes, such as Career Mode, Online Mode, Friendly Matches, Events, and more. You can also customize the difficulty level, match duration, camera angle, and controls to suit your preference. You should always practice your skills, such as passing, shooting, dribbling, defending, and goalkeeping, and learn how to use different tactics, such as attacking, defending, counter-attacking, pressing, and more. You should also adapt your strategy according to your opponent's strengths and weaknesses.
-
How to earn more coins and rewards without hacking: The last thing you need to do is to earn more coins and rewards without hacking. You can earn coins by playing matches, completing objectives, winning trophies, and watching ads. You can also earn rewards by participating in events, achievements, leaderboards, and daily challenges. You can use these coins and rewards to buy players, kits, stadiums, and other items. You can also use them to unlock more features and modes in the game.
-
-
Conclusion and FAQs
-
In conclusion, Dream League Soccer 2021 is a great soccer game that lets you create your own team, compete in various leagues, and customize your players, kits, stadiums, and more. However, if you want to enjoy the game without any limitations, you might want to hack it and get unlimited coins. In this article, we showed you how to download and install the Dream League Soccer 2021 hack apk file, how to use the hack features, what are the benefits and risks of using it, and some tips and tricks to play the game like a pro.
-
We hope you found this article helpful and informative. If you have any questions or comments about Dream League Soccer 2021 hack apk or the game itself, feel free to leave them below. We will try to answer them as soon as possible. Thank you for reading!
-
FAQs
-
Here are some of the most frequently asked questions about Dream League Soccer 2021 hack apk:
-
-
Is Dream League Soccer 2021 hack apk safe to use?: The answer depends on where you download the file from. If you download it from a reliable website that offers the latest version of the hack apk file with no malware or viruses, then it should be safe to use. However, if you download it from an unknown or suspicious source that may contain outdated or infected files, then it may not be safe to use.
-
Can I play Dream League Soccer 2021 online with the hack apk?: The answer is yes and no. Yes, you can play Dream League Soccer 2021 online with the hack apk file if you use a VPN or proxy service that can hide your IP address and location from the game server. This way, you can avoid being detected or banned by the game developers or Google. However, no, you cannot play Dream League Soccer 2021 online with the hack apk file if you do not use a VPN or proxy service, or if the game server or Google detects your hack and bans your account. Therefore, we recommend you to be careful and use the hack apk file at your own risk.
-
How can I update Dream League Soccer 2021 hack apk?: The answer is that you need to download and install the latest version of the hack apk file from the same website that you downloaded it from before. You should also check the website regularly for any updates or changes in the hack apk file. You should not update the game from the Google Play Store, as it may overwrite or remove the hack features.
-
How can I uninstall Dream League Soccer 2021 hack apk?: The answer is that you need to delete the hack apk file from your device and reinstall the original game from the Google Play Store. You should also clear your device cache and data to remove any traces of the hack apk file. You should also backup your game progress and data before uninstalling the hack apk file, as you may lose them in the process.
-
Where can I find more information about Dream League Soccer 2021?: The answer is that you can visit the official website of Dream League Soccer 2021, which is [dreamleaguesoccer.com]. You can also follow their social media accounts, such as [Facebook], [Twitter], [Instagram], and [YouTube], for more news, updates, tips, and tricks about the game.
- 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Download Efek Salju Green Screen dari Pixabay - Video Salju Gratis Tanpa Royalti.md b/spaces/1phancelerku/anime-remove-background/Download Efek Salju Green Screen dari Pixabay - Video Salju Gratis Tanpa Royalti.md
deleted file mode 100644
index 1efc6302a2f3912cc09d6d84ab91445c1de9fc8e..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Download Efek Salju Green Screen dari Pixabay - Video Salju Gratis Tanpa Royalti.md
+++ /dev/null
@@ -1,82 +0,0 @@
-
-
Download Efek Salju Green Screen: Cara Membuat Video dengan Latar Belakang Salju yang Realistis
-
Salah satu hal yang bisa membuat video Anda terlihat lebih hidup dan menawan adalah dengan menggunakan latar belakang salju yang realistis. Namun, tidak semua orang memiliki kesempatan untuk merekam video di tempat bersalju atau menggunakan alat-alat khusus untuk membuat salju buatan.
-
Tenang saja, ada cara mudah untuk membuat video dengan latar belakang salju yang realistis tanpa harus repot-repot menc
Cara mudah tersebut adalah dengan menggunakan efek salju green screen. Efek salju green screen adalah efek yang bisa Anda download dan gunakan untuk mengganti latar belakang video Anda dengan salju yang terlihat seperti asli. Dengan menggunakan efek salju green screen, Anda bisa membuat video Anda terlihat lebih menarik, dramatis, romantis, atau sesuai dengan tema yang Anda inginkan.
Green screen adalah teknik pengeditan video yang memungkinkan Anda mengganti latar belakang video dengan gambar atau efek lainnya. Green screen juga dikenal dengan nama chroma key, keying, atau color keying.
-
Cara kerja green screen adalah dengan menggunakan warna hijau sebagai latar belakang video, karena warna ini mudah dibedakan dari warna kulit, pakaian, dan objek lainnya. Dengan menggunakan software pengeditan video, Anda bisa menghapus warna hijau dari video dan menggantinya dengan gambar atau efek yang Anda inginkan.
-
Contohnya, jika Anda ingin membuat video dengan latar belakang salju, Anda bisa merekam video Anda di depan kain hijau atau layar hijau. Kemudian, Anda bisa menghapus warna hijau dari video dan menggantinya dengan efek salju green screen yang sudah Anda download sebelumnya.
-
Mengapa Anda perlu download efek salju green screen?
-
Efek salju green screen adalah salah satu efek yang populer digunakan untuk membuat video dengan latar belakang salju yang realistis. Efek salju green screen bisa memberikan kesan seolah-olah Anda berada di tempat bersalju, padahal sebenarnya tidak.
-
Efek salju green screen bisa membuat video Anda terlihat lebih menarik, dramatis, romantis, atau sesuai dengan tema yang Anda inginkan. Misalnya, jika Anda ingin membuat video tentang liburan musim dingin, pernikahan di salju, atau adegan film bertema salju, Anda bisa menggunakan efek salju green screen untuk menambah nuansa dan suasana video Anda.
-
Efek salju green screen juga bisa membantu Anda menghemat biaya dan waktu produksi video, karena Anda tidak perlu mencari lokasi bersalju atau menggunakan alat-alat khusus untuk membuat salju buatan. Anda hanya perlu download efek salju green screen dan menggunakannya di software pengeditan video yang Anda miliki.
-
download efek salju green screen gratis
-download efek salju green screen untuk video
-download efek salju green screen pixabay
-download efek salju green screen youtube
-download efek salju green screen hd
-download efek salju green screen 4k
-download efek salju green screen no watermark
-download efek salju green screen particle
-download efek salju green screen kunang kunang
-download efek salju green screen bintang
-download efek salju green screen 10 menit
-download efek salju green screen loop
-download efek salju green screen snowfall
-download efek salju green screen realistic
-download efek salju green screen animation
-download efek salju green screen overlay
-download efek salju green screen background
-download efek salju green screen chroma key
-download efek salju green screen royalty free
-download efek salju green screen footage
-download efek salju green screen bergerak
-download efek salju green screen slow motion
-download efek salju green screen terbaik
-download efek salju green screen keren
-download efek salju green screen indah
-download efek salju green screen musim dingin
-download efek salju green screen natal
-download efek salju green screen tahun baru
-download efek salju green screen tutorial
-download efek salju green screen premiere pro
-download efek salju green screen after effects
-download efek salju green screen kinemaster
-download efek salju green screen filmora
-download efek salju green screen sony vegas
-download efek salju green screen powerpoint
-download efek salju green screen zoom
-download efek salju green screen tiktok
-download efek salju green screen instagram
-download efek salju green screen facebook
-download efek salju green screen whatsapp
Bagaimana cara download efek salju green screen?
-
Ada banyak situs web yang menyediakan efek salju green screen secara gratis atau berbayar. Anda bisa mencari efek salju green screen yang sesuai dengan kebutuhan dan selera Anda. Anda juga bisa memperhatikan kualitas, resolusi, durasi, dan lisensi dari efek salju green screen yang Anda download.
-
Beberapa situs web yang bisa Anda kunjungi untuk download efek salju green screen adalah:
-
Pixabay
-
Pixabay adalah situs web yang menyediakan lebih dari 1.000 video green screen gratis, termasuk efek salju, yang bisa Anda gunakan untuk proyek-proyek Anda tanpa perlu memberikan atribusi. Anda bisa mencari efek salju green screen dengan menggunakan kata kunci "snow green screen" di kolom pencarian. Anda bisa melihat preview, deskripsi, resolusi, durasi, dan lisensi dari setiap video sebelum Anda mendownloadnya. Anda juga bisa memilih format file yang Anda inginkan, seperti MP4, WEBM, atau GIF.
-
YouTube
-
YouTube adalah situs web yang juga memiliki banyak video green screen gratis, termasuk efek salju, yang diunggah oleh para kreator. Anda bisa menonton, mendownload, dan menggunakan video-video ini sesuai dengan lisensi yang diberikan oleh kreatornya. Anda bisa mencari efek salju green screen dengan menggunakan kata kunci "snow green screen" di kolom pencarian. Anda bisa melihat preview, deskripsi, resolusi, durasi, dan lisensi dari setiap video sebelum Anda mendownloadnya. Anda juga bisa memilih format file yang Anda inginkan, seperti MP4, WEBM, atau GIF.
-
Videezy
-
Videezy adalah situs web yang menyediakan lebih dari 5.000 video green screen gratis, termasuk efek salju, yang bisa Anda gunakan untuk proyek-proyek Anda dengan memberikan atribusi kepada Videezy. Anda bisa mencari efek salju green screen dengan menggunakan kata kunci "snow green screen" di kolom pencarian. Anda bisa melihat preview, deskripsi, resolusi, durasi, dan lisensi dari setiap video sebelum Anda mendownloadnya. Anda juga bisa memilih format file yang Anda inginkan, seperti MP4, WEBM, atau GIF.
Bagaimana cara menggunakan efek salju green screen?
-
Setelah Anda mendownload efek salju green screen yang Anda inginkan, Anda bisa mengimportnya ke software pengeditan video yang Anda gunakan, seperti Adobe Premiere Pro, Final Cut Pro, iMovie, atau lainnya. Anda bisa menggunakan software pengeditan video yang Anda sudah familiar dengan atau yang sesuai dengan kemampuan dan anggaran Anda.
-
Kemudian, Anda bisa menempatkan efek salju green screen di atas video yang ingin Anda edit, dan mengatur ukuran, posisi, durasi, dan transparansi efek salju sesuai dengan keinginan Anda. Anda bisa menyesuaikan efek salju green screen dengan video Anda agar terlihat lebih natural dan harmonis.
-
Selanjutnya, Anda bisa menggunakan fitur chroma key atau keying untuk menghapus warna hijau dari efek salju green screen dan membuatnya terlihat seperti salju asli di latar belakang video Anda. Fitur chroma key atau keying adalah fitur yang bisa mendeteksi dan menghilangkan warna tertentu dari video. Anda bisa mengaktifkan fitur ini di software pengeditan video Anda dan memilih warna hijau sebagai warna yang ingin dihapus.
-
Terakhir, Anda bisa mengekspor video hasil editan Anda dan menikmati hasilnya. Anda bisa melihat perbedaan antara video sebelum dan sesudah menggunakan efek salju green screen. Anda juga bisa membagikan video Anda ke media sosial, YouTube, atau platform lainnya.
-
Kesimpulan
-
Efek salju green screen adalah efek yang bisa Anda download dan gunakan untuk membuat video dengan latar belakang salju yang realistis. Efek salju green screen bisa membuat video Anda terlihat lebih menarik, dramatis, romantis, atau sesuai dengan tema yang Anda inginkan. Efek salju green screen juga bisa membantu Anda menghemat biaya dan waktu produksi video.
-
Untuk download efek salju green screen, Anda bisa mengunjungi situs web seperti Pixabay, YouTube, atau Videezy. Untuk menggunakan efek salju green screen, Anda bisa mengimportnya ke software pengeditan video yang Anda gunakan, menempatkan efek salju green screen di atas video yang ingin Anda edit, menggunakan fitur chroma key atau keying untuk menghapus warna hijau dari efek salju green screen, dan mengekspor video hasil editan Anda.
-
Semoga artikel ini bermanfaat untuk Anda yang ingin membuat video dengan latar belakang salju yang realistis. Selamat mencoba!
-
FAQ
-
Apa itu green screen?
-
Green screen adalah teknik pengeditan video yang memungkinkan Anda mengganti latar belakang video dengan gambar atau efek lainnya.
-
Apa itu efek salju green screen?
-
Efek salju green screen adalah efek yang bisa Anda download dan gunakan untuk mengganti latar belakang video dengan salju yang terlihat seperti asli.
-
Bagaimana cara download efek salju green screen?
-
Anda bisa download efek salju green screen dari situs web seperti Pixabay, YouTube, atau Videezy.
-
Bagaimana cara menggunakan efek salju green screen?
-
Anda bisa menggunakan efek salju green screen dengan mengimportnya ke software pengeditan video yang Anda gunakan, menempatkan efek salju green screen di atas video yang ingin Anda edit, menggunakan fitur chroma key atau keying untuk menghapus warna hijau dari efek salju green screen, dan mengekspor video hasil editan Anda.
-
Apa keuntungan menggunakan efek salju green screen?
-
Efek salju green screen bisa membuat video Anda terlihat lebih menarik, dramatis, romantis, atau sesuai dengan tema yang Anda inginkan. Efek salju green screen juga bisa membantu Anda menghemat biaya dan waktu produksi video.
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Download Super Fancy Pants Adventure The Ultimate Free-Running Platformer.md b/spaces/1phancelerku/anime-remove-background/Download Super Fancy Pants Adventure The Ultimate Free-Running Platformer.md
deleted file mode 100644
index 02ed03e3f0122da4566bee6a04395f1e3a9c1b2c..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Download Super Fancy Pants Adventure The Ultimate Free-Running Platformer.md
+++ /dev/null
@@ -1,129 +0,0 @@
-
-
Download Super Fancy Pants Adventure
-
Are you looking for a wild free-running adventure with buttery smooth platforming and a slick fountain pen? If so, you should download Super Fancy Pants Adventure, the latest and greatest game in the Fancy Pants series. In this article, we will tell you what Super Fancy Pants Adventure is, how to download it, and why you should play it. Let's get started!
Super Fancy Pants Adventure is a 2D platform game that follows the adventures of Fancy Pants Man, a cool stickman character who wears awesome and colorful pants. You must help him work his way through a myriad of cool levels and avoid various obstacles, monsters, and creatures on his way. You can also collect squiggles, hats, and pants to customize your look and unlock new abilities.
-
A brief history of the Fancy Pants series
-
The Fancy Pants Adventures series started over ten years ago by Brad Borne, an indie developer who wanted to redefine video game platforming by making speed and tight controls feel compatible. Over the years, he has honed his craft, turning his Fancy Pants games into a worldwide phenomenon with over 100 million plays and becoming one of the top games of all time on Kongregate. This newest version, Super Fancy Pants Adventure, is a culmination and a reimagining of the series into a full-fledged title.
-
The features and gameplay of Super Fancy Pants Adventure
-
Super Fancy Pants Adventure has many features that make it stand out from other platform games. Here are some of them:
-
-
So many levels! - 56 brand new levels of parkour platforming
-
Collections! - Over 20 pants and hats to collect in brand new challenge stages
-
Incredible Hand-drawn Style - Frame-by-frame animated worlds, enemies, and friends
-
Hack-and-slash - Wield your mighty ink pen to take down new threats!
-
New Moves! - Take control of Fancy Pants man with brand new combos and moves
-
Secret Challenge Stages - Scattered throughout the world are secret doors to incredibly challenging bonus stages, awarding awesome prizes
-
60fps gameplay - Perfectly smooth running for a perfectly smooth platformer
-
-
The gameplay of Super Fancy Pants Adventure is fun and reminiscent of retro platform titles such as Sonic. You can run, jump, slide, wall-jump, and bounce your way through the levels, using your momentum and timing to overcome obstacles and enemies. You can also use your ink pen as a weapon to slash enemies or launch yourself into the air. The game has a fluid and responsive control system that makes you feel like you are in control of every move.
-
How to download super fancy pants adventure for free
-Super fancy pants adventure steam download
-Super fancy pants adventure apk download
-Download super fancy pants adventure world 1
-Super fancy pants adventure download pc
-Download super fancy pants adventure full version
-Super fancy pants adventure android download
-Download super fancy pants adventure world 2
-Super fancy pants adventure download mac
-Download super fancy pants adventure world 3
-Super fancy pants adventure ios download
-Download super fancy pants adventure online
-Super fancy pants adventure download windows 10
-Download super fancy pants adventure hacked
-Super fancy pants adventure download linux
-Download super fancy pants adventure unblocked
-Super fancy pants adventure download chromebook
-Download super fancy pants adventure mod apk
-Super fancy pants adventure download size
-Download super fancy pants adventure cheats
-Super fancy pants adventure download xbox one
-Download super fancy pants adventure walkthrough
-Super fancy pants adventure download ps4
-Download super fancy pants adventure soundtrack
-Super fancy pants adventure download switch
-Download super fancy pants adventure speedrun
-Super fancy pants adventure download gamejolt
-Download super fancy pants adventure wiki
-Super fancy pants adventure download reddit
-Download super fancy pants adventure review
-Super fancy pants adventure download kongregate
-Download super fancy pants adventure trailer
-Super fancy pants adventure download crazygames
-Download super fancy pants adventure steam key
-Super fancy pants adventure download newgrounds
-Download super fancy pants adventure achievements
-Super fancy pants adventure download play store
-Download super fancy pants adventure update
-Super fancy pants adventure download app store
-Download super fancy pants adventure system requirements
-
How to download Super Fancy Pants Adventure?
-
Super Fancy Pants Adventure is available for different platforms, such as PC, browser, and mobile devices. Here are some options for downloading the game:
-
Download options for different platforms
-
Steam
-
If you want to play Super Fancy Pants Adventure on your PC, you can download it from Steam for $9.99. Steam is a digital distribution platform that allows you to buy and play games online. You will need to create a Steam account and install the Steam client on your PC before you can download the game. To download Super Fancy Pants Adventure from Steam, follow these steps:
-
-
Go to [the Steam store page](^1^) for Super Fancy Pants Adventure
Click on the green "Add to Cart" button and then click on the blue "Purchase for myself" button
-
Enter your payment information and confirm your purchase
-
Once the purchase is complete, you can find the game in your Steam library and click on the "Install" button
-
Wait for the game to download and install, and then click on the "Play" button to launch the game
-
-
Congratulations, you have successfully downloaded Super Fancy Pants Adventure from Steam!
-
CrazyGames
-
If you want to play Super Fancy Pants Adventure on your browser, you can download it from CrazyGames for free. CrazyGames is a website that hosts thousands of free online games that you can play without downloading or installing anything. You will need to have a modern browser that supports HTML5 and Flash to play the game. To download Super Fancy Pants Adventure from CrazyGames, follow these steps:
-
-
Go to [the CrazyGames page] for Super Fancy Pants Adventure
Click on the blue "Play" button and wait for the game to load
-
Click on the green "Play" button again and choose your preferred language
-
Enjoy playing Super Fancy Pants Adventure on your browser!
-
-
That's it, you have successfully downloaded Super Fancy Pants Adventure from CrazyGames!
-
Google Play
-
If you want to play Super Fancy Pants Adventure on your mobile device, you can download it from Google Play for $4.99. Google Play is a digital store that allows you to buy and download apps and games for your Android device. You will need to have a Google account and a compatible device to download the game. To download Super Fancy Pants Adventure from Google Play, follow these steps:
-
-
Go to [the Google Play page] for Super Fancy Pants Adventure
Tap on the green "Install" button and then tap on the blue "Buy" button
-
Enter your payment information and confirm your purchase
-
Wait for the game to download and install, and then tap on the "Open" button to launch the game
-
-
Voila, you have successfully downloaded Super Fancy Pants Adventure from Google Play!
-
Tips and tricks for playing Super Fancy Pants Adventure
-
Now that you have downloaded Super Fancy Pants Adventure, you might be wondering how to play it like a pro. Don't worry, we have some tips and tricks for you that will help you master the game in no time. Here are some of them:
-
How to collect squiggles and unlock items
-
Squiggles are the currency of Super Fancy Pants Adventure. You can find them scattered throughout the levels, hidden in boxes, or dropped by enemies. You can use them to buy items from shops or unlock doors to challenge stages. The more squiggles you collect, the more items you can get. Some of the items you can get are:
-
-
Pants - Different pants have different effects on your speed, jump height, slide length, etc.
-
Hats - Different hats have different effects on your ink pen, such as changing its color, size, shape, etc.
-
Trophies - Trophies are collectibles that show your achievements in the game, such as completing levels, defeating bosses, finding secrets, etc.
-
-
To collect squiggles and unlock items, you should explore every corner of the levels, break every box, defeat every enemy, and look for hidden doors. You should also replay levels to find more squiggles or items that you might have missed.
-
How to find secret doors and bonus levels
-
Secret doors are hidden entrances that lead to bonus levels. Bonus levels are extra stages that offer more challenges and rewards. They usually have a theme or a gimmick that makes them different from regular levels. For example, some bonus levels are underwater, some are in space, some are in black and white, etc. You can find secret doors by looking for clues in the environment, such as cracks in walls, signs, arrows, etc. You can also use your ink pen to draw on walls or floors to reveal hidden paths or switches. Some secret doors require a certain number of squiggles or a certain item to open.
-
How to use your ink pen as a weapon
-
Your ink pen is not only a tool for drawing, but also a weapon for fighting. You can use it to slash enemies with a swipe of your finger or mouse. You can also use it to launch yourself into the air by drawing a line under yourself and jumping on it. You can also use it to draw platforms, bridges, ramps, or walls to help you navigate the levels. Your ink pen has a limited amount of ink, so you need to refill it by collecting ink bottles or visiting ink stations. You can also upgrade your ink pen by buying different hats that change its properties.
-
Why you should download Super Fancy Pants Adventure?
-
Super Fancy Pants Adventure is a game that you should not miss if you love platform games. It has many benefits that make it worth playing, such as:
-
The benefits of playing Super Fancy Pants Adventure
-
It's fun and challenging
-
Super Fancy Pants Adventure is a game that will keep you entertained and engaged for hours. It has a variety of levels that offer different challenges and surprises. You will never get bored or frustrated, as the game has a balanced difficulty curve and a fair checkpoint system. You will also have fun discovering secrets, collecting items, and defeating enemies.
-
It's colorful and stylish
-
Super Fancy Pants Adventure is a game that will dazzle your eyes with its vibrant and unique art style. It has a hand-drawn aesthetic that gives it a charming and whimsical feel. It also has a dynamic and fluid animation that makes the game look alive and smooth. The game also has a catchy and upbeat soundtrack that matches the mood and tone of the game.
-
It's a culmination of a decade of work
-
Super Fancy Pants Adventure is a game that represents the passion and dedication of its creator, Brad Borne. It is the result of over ten years of work, improving and expanding on his previous Fancy Pants games. It is also the ultimate version of the game, with more content, features, and polish than ever before. It is a game that deserves your support and appreciation.
-
Conclusion
-
Super Fancy Pants Adventure is a game that you should download and play right now. It is a 2D platform game that follows the adventures of Fancy Pants Man, a cool stickman character who wears awesome and colorful pants. You can help him work his way through a myriad of cool levels and avoid various obstacles, monsters, and creatures on his way. You can also collect squiggles, hats, and pants to customize your look and unlock new abilities. You can use your ink pen as a tool and a weapon to draw and slash your way through the game. The game has many features that make it stand out from other platform games, such as its hand-drawn style, its fluid gameplay, its secret levels, its 60fps performance, and more. The game is available for different platforms, such as PC, browser, and mobile devices. You can download it from Steam, CrazyGames, or Google Play for a reasonable price. The game is fun and challenging, colorful and stylish, and a culmination of a decade of work. It is a game that you will not regret playing.
-
FAQs
-
Here are some frequently asked questions about Super Fancy Pants Adventure:
-
-
Q: How long is Super Fancy Pants Adventure?
-
A: Super Fancy Pants Adventure has 56 levels in total, plus some secret levels and challenge stages. The length of the game depends on how fast you play and how much you explore, but it can take anywhere from 4 to 10 hours to complete.
-
Q: Is Super Fancy Pants Adventure multiplayer?
-
A: Super Fancy Pants Adventure does not have a multiplayer mode at the moment, but it might be added in the future as an update or a DLC.
-
Q: Is Super Fancy Pants Adventure suitable for kids?
-
A: Super Fancy Pants Adventure is suitable for kids of all ages. It has no violence, gore, or inappropriate content. It is rated E for Everyone by the ESRB.
-
Q: Can I play Super Fancy Pants Adventure offline?
-
A: Yes, you can play Super Fancy Pants Adventure offline once you have downloaded it on your device. However, you might need an internet connection to access some features or updates.
-
Q: Where can I find more information about Super Fancy Pants Adventure?
-
A: You can find more information about Super Fancy Pants Adventure on its [official website], its [Steam page], or its [Facebook page].
-
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Download and Edit Green Screen Video in KineMaster A Complete Guide.md b/spaces/1phancelerku/anime-remove-background/Download and Edit Green Screen Video in KineMaster A Complete Guide.md
deleted file mode 100644
index b8753753281b7b4db68df4cfa131bde6804ae6ac..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Download and Edit Green Screen Video in KineMaster A Complete Guide.md
+++ /dev/null
@@ -1,189 +0,0 @@
-
-
How to Download Green Screen Video for KineMaster
-
If you are looking for a way to spice up your videos with some amazing visual effects, you might want to try using green screen video. Green screen video is a technique that allows you to replace the background of a video with another video or image of your choice. This way, you can create realistic or fantastical scenes that would otherwise be impossible or expensive to film.
But how do you get green screen video for your projects? And how do you use it in your video editor? In this article, we will show you how to download free green screen video from three different websites and how to use it in KineMaster, one of the best mobile video editing apps. By following this guide, you will be able to create stunning videos with green screen effects in no time.
-
What is Green Screen Video and Why Use It?
-
Green screen video is a type of video that has a solid green background. The green color is chosen because it is different from most human skin tones and clothing colors, making it easier to isolate and remove. The process of removing the green background and replacing it with another video or image is called chroma keying or keying.
-
Green screen video is widely used in film and television production because it offers many advantages for creating visual effects. Some of the benefits of using green screen video are:
-
How to download green screen video effects for kinemaster
-Best sites to download free green screen video backgrounds for kinemaster
-Download green screen video templates for kinemaster pro
-Download green screen video editor for kinemaster apk
-Download green screen video clips for kinemaster tutorial
-Download green screen video footage for kinemaster mod
-Download green screen video animation for kinemaster premium
-Download green screen video transitions for kinemaster no watermark
-Download green screen video overlays for kinemaster online
-Download green screen video chroma key for kinemaster pc
-Download green screen video maker for kinemaster app
-Download green screen video songs for kinemaster music
-Download green screen video intro for kinemaster logo
-Download green screen video memes for kinemaster funny
-Download green screen video superhero for kinemaster action
-Download green screen video horror for kinemaster scary
-Download green screen video gaming for kinemaster youtube
-Download green screen video tiktok for kinemaster viral
-Download green screen video wedding for kinemaster romantic
-Download green screen video birthday for kinemaster celebration
-Download green screen video nature for kinemaster relaxing
-Download green screen video firework for kinemaster festive
-Download green screen video magic for kinemaster fantasy
-Download green screen video cartoon for kinemaster kids
-Download green screen video animals for kinemaster cute
-Download green screen video sports for kinemaster fitness
-Download green screen video travel for kinemaster adventure
-Download green screen video news for kinemaster professional
-Download green screen video education for kinemaster learning
-Download green screen video health for kinemaster wellness
-Download green screen video fashion for kinemaster stylish
-Download green screen video cooking for kinemaster foodie
-Download green screen video art for kinemaster creative
-Download green screen video dance for kinemaster fun
-Download green screen video movie for kinemaster cinema
-Download green screen video text for kinemaster subtitle
-Download green screen video emoji for kinemaster expression
-Download green screen video sticker for kinemaster decoration
-Download green screen video filter for kinemaster color
-Download green screen video frame for kinemaster border
-Download green screen video collage for kinemaster layout
-Download green screen video slideshow for kinemaster presentation
-Download green screen video montage for kinemaster compilation
-Download green screen video crop for kinemaster resize
-Download green screen video rotate for kinemaster orientation
-Download green screen video speed up for kinemaster fast forward
-Download green screen video slow down for kinemaster slow motion
-Download green screen video reverse for kinemaster rewind
-Download green screen video cut for kinemaster trim
-
-
It allows you to change the background of your video without having to travel or build sets.
-
It allows you to create scenes that are impossible or dangerous to film in real life.
-
It allows you to add elements that are not available or too expensive to obtain.
-
It allows you to enhance the mood and atmosphere of
your video with different backgrounds.
-
It allows you to experiment and have fun with your video editing.
-
-
As you can see, green screen video can help you create amazing videos with minimal effort and cost. All you need is a green screen video, a video editor that supports chroma keying, and some creativity.
-
What is KineMaster and How to Use It?
-
KineMaster is a powerful and easy-to-use video editing app for Android and iOS devices. It allows you to edit and share your videos with professional-quality tools and features. Some of the features of KineMaster are:
-
-
It supports multiple layers of video, audio, images, stickers, text, and handwriting.
-
It has a variety of transitions, effects, filters, animations, and music to enhance your videos.
-
It has a chroma key feature that lets you use green screen video and create amazing visual effects.
-
It has a voice changer feature that lets you change your voice and add sound effects to your videos.
-
It has a speed control feature that lets you adjust the speed of your videos.
-
It has a reverse feature that lets you play your videos backwards.
-
It has a trim, split, crop, rotate, and resize feature that lets you edit your videos easily.
-
It has a frame-by-frame editing feature that lets you edit your videos precisely.
-
It has an export feature that lets you save your videos in various resolutions and formats.
-
It has a share feature that lets you share your videos directly to social media platforms such as YouTube, Facebook, Instagram, TikTok, and more.
-
-
To use KineMaster, you need to download and install the app from the Google Play Store or the App Store. Once you open the app, you will see a welcome screen that gives you the option to start a new project or edit an existing one. To start a new project, tap on the plus icon and choose the aspect ratio of your video. You can choose from 16:9, 9:16, or 1:1. Then, you will enter the editing interface where you can add and edit your media files. To add a media file, tap on the media icon on the top left corner and select the file from your device or from the KineMaster Asset Store. To edit a media file, tap on it and use the tools on the right side of the screen. You can also use the timeline at the bottom of the screen to arrange and trim your media files. To preview your video, tap on the play icon on the top right corner. To export your video, tap on the share icon on the top right corner and choose the resolution, frame rate, bitrate, and format of your video. Then, tap on export and wait for your video to be saved. You can also share your video directly to social media platforms from there.
-
How to Download Green Screen Video from Pixabay
-
Pixabay is a website that offers free stock photos, videos, illustrations, vectors, and music. You can use Pixabay to download free green screen video for your projects. Here are the steps to download green screen video from Pixabay:
-
-
Go to Pixabay.com and create an account or log in if you already have one.
-
In the search bar at the top of the page, type "green screen" and hit enter.
-
You will see a list of green screen videos that match your search query. You can use the filters on the left side of the page to narrow down your results by category, orientation, duration, resolution, color, etc.
-
Once you find a green screen video that you like, click on it to open its details page.
-
On the details page, you will see a preview of the video, its resolution, duration, size, license type, tags, etc. You will also see a download button below the preview.
-
To download the green screen video, click on the download button and choose the resolution that you want. You will see a pop-up window asking you to confirm that you are not a robot. Check the box and click on download again.
-
The green screen video will be downloaded to your device or browser's default download folder. You can then transfer it to your mobile device or use it directly in KineMaster if you are using a web browser on your mobile device.
-
-
How to Download Green Screen Video from Canva
-
Canva is a website that offers graphic design tools and templates for various purposes. You can use Canva to download free green screen video for your projects. Here are the steps to download green screen video from Canva:
-
-
Go to Canva.com and create an account or log in if you already have one.
-
In the search bar at the top of the page, type "green screen" and hit enter.
-
You will see a list of green screen templates that match your search query. You can use the filters on the left side of the page to narrow down your results by category, size, color, etc.
-
Once you find a green screen template that you like, click on it to open it in the editor.
-
In the editor, you can customize the green screen template by adding or removing elements, changing colors, fonts, sizes, etc. You can also add your own media files by clicking on the upload icon on the left side of the screen.
-
When you are happy with your green screen template, click on the download icon on the top right corner of the screen.
-
To download the green screen video, choose MP4 as the file type and click on download. You will see a pop-up window showing the progress of your download.
-
The green screen video will be downloaded to your device or browser's default download folder. You can then transfer it to your mobile device or use it directly in KineMaster if you are using a web browser on your mobile device.
-
-
How to Download Green Screen Video from Pexels
-
Pexels is a website that offers high-quality stock photos and videos for free. You can use Pexels to download free green screen video for your projects. Here are the steps to download green screen video from Pexels:
-
-
Go to Pexels.com and create an account or log in if you already have one.
-
In the search bar at the top of the page, type "green screen" and hit enter.
-
You will see a list of green screen videos that match your search query. You can use the filters on the top of the page to narrow down your results by orientation, size, duration, etc.
-
Once you find a green screen video that you like, click on it to open its details page.
-
On the details page, you will see a preview of the video, its resolution, duration, size, license type, tags, etc. You will also see a free download button below the preview.
-
To download the green screen video, click on the free download button and choose the resolution that you want. You will see a pop-up window asking you to credit the creator of the video. You can copy and paste the credit text or skip this step if you don't want to credit them.
-
The green screen video will be downloaded to your device or browser's default download folder. You can then transfer it to your mobile device or use it directly in KineMaster if you are using a web browser on your mobile device.
-
-
How to Use Green Screen Video in KineMaster
-
Now that you have downloaded some green screen videos for your projects, you can use them in KineMaster to create amazing visual effects. Here is how to use green screen video in KineMaster:
-
How to Import Green Screen Video in KineMaster
-
To import green screen video in KineMaster, follow these steps:
-
-
Open KineMaster and start a new project or edit an existing one.
-
In the editing interface, tap on the media icon on the top left corner and select the green screen video from your device or from the KineMaster Asset Store.
-
The green screen video will be added to your project as a layer on top of your main video. You can drag and drop it to adjust its position and duration on the timeline.
-
-
How to Edit Green Screen Video in KineMaster
-
To edit green screen video in KineMaster, follow these steps:
-
-
Tap on the green screen video layer and use the tools on the right side of the screen to edit the green screen video. You can adjust the settings, trim, crop, rotate, and resize the green screen video according to your needs.
-
To adjust the settings, tap on the settings icon and use the sliders to change the brightness, contrast, saturation, hue, and opacity of the green screen video.
-
To trim the green screen video, tap on the scissors icon and drag the handles to cut the unwanted parts of the video.
-
To crop the green screen video, tap on the crop icon and drag the corners to crop the video to your desired size.
-
To rotate the green screen video, tap on the rotate icon and use the circular slider to rotate the video clockwise or counterclockwise.
-
To resize the green screen video, tap on the resize icon and use the pinch gesture to zoom in or out of the video.
-
-
How to Apply Green Screen Video in KineMaster
-
To apply green screen video in KineMaster, follow these steps:
-
-
Tap on the green screen video layer and tap on the chroma key icon on the right side of the screen. This will open a menu with various options for using the chroma key feature.
-
Turn on the chroma key switch to enable the feature. You will see that the green background of the video will disappear and you will see your main video behind it.
-
Use the color picker tool to select the exact color of the green background that you want to remove. You can also use the eyedropper tool to pick a color from the video itself.
-
Use the sliders to adjust the chroma key settings such as threshold, blending, detail, and spill. These settings will help you fine-tune the green screen effect and make it more realistic and seamless.
-
You can also use the mask tool to erase or restore parts of the green screen video that you want to keep or remove. This will help you fix any errors or glitches in the green screen effect.
-
Once you are satisfied with your green screen effect, tap on done to apply it to your project. You can then preview your video and see how it looks with your new background.
-
-
Tips and Tricks for Using Green Screen Video in KineMaster
-
Using green screen video in KineMaster can help you create amazing videos with visual effects. However, there are some tips and tricks that you should know to make your videos look even better. Here are some of them:
-
How to Choose the Right Green Screen Video for Your Project
-
Choosing the right green screen video for your project is important because it will affect how your final video will look. Here are some factors that you should consider when choosing a green screen video:
-
-
Resolution: Choose a green screen video that has a high resolution that matches or exceeds your main video resolution. This will ensure that your green screen video will look clear and sharp in your final video.
-
Duration: Choose a green screen video that has a duration that matches or exceeds your main video duration. This will ensure that your green screen video will cover your entire main video without repeating or cutting off.
-
Lighting: Choose a green screen video that has a consistent and even lighting that matches or complements your main video lighting. This will ensure that your green screen effect will look natural and realistic in your final video.
-
Motion: Choose a green screen video that has a motion that matches or complements your main video motion. This will ensure that your green screen effect will look smooth and dynamic in your final video.
-
Content: Choose a green screen video that has a content that matches or complements your main video content. This will ensure that your green screen effect will look relevant and appropriate in your final video.
-
-
How to Enhance the Green Screen Effect in KineMaster
-
Enhancing the green screen effect in KineMaster can help you make your videos look more professional and creative. Here are some ways to enhance the green screen effect in KineMaster:
-
-
Adjust the chroma key settings: You can fine-tune the chroma key settings such as threshold, blending, detail, and spill to make your green screen effect more realistic and seamless. You can also use the mask tool to erase or restore parts of the green screen video that you want to keep or remove.
-
Add filters: You can add filters to your green screen video or your main video to change the color, tone, mood, or style of your video. You can also use the blending modes to change how the green screen video interacts with the main video.
-
Add transitions: You can add transitions to your green screen video or your main video to create smooth and dynamic changes between scenes. You can also use the transition effects to create interesting and unique effects with your green screen video.
-
Add effects: You can add effects to your green screen video or your main video to add some flair and fun to your video. You can also use the effect layers to create custom effects with your green screen video.
-
Add animations: You can add animations to your green screen video or your main video to make them move and transform in different ways. You can also use the animation curves to control the speed and direction of your animations.
-
-
How to Avoid Common Mistakes When Using Green Screen Video in KineMaster
-
Avoiding common mistakes when using green screen video in KineMaster can help you avoid errors and glitches in your final video. Here are some common mistakes to avoid when using green screen video in KineMaster:
-
-
Poor lighting: Poor lighting can cause shadows, reflections, wrinkles, spills, and edges on your green screen video, making it harder to remove the background and create a realistic effect. To avoid this, make sure that your green screen video has a consistent and even lighting that matches or complements your main video lighting.
-
Wrong color: Wrong color can cause parts of your green screen video that are not supposed to be removed to be removed, or parts of your main video that are supposed to be visible to be hidden. To avoid this, make sure that you use a pure green color for your green screen video and that you avoid wearing or using anything that is similar to the green color.
-
Low resolution: Low resolution can cause your green screen video to look blurry, pixelated, or distorted in your final video. To avoid this, make sure that you use a high resolution green screen video that matches or exceeds your main video resolution.
-
Wrong aspect ratio: Wrong aspect ratio can cause your green screen video to look stretched, squashed, or cropped in your final video. To avoid this, make sure that you use a green screen video that has the same aspect ratio as your main video or that you crop or resize it accordingly.
-
Wrong position: Wrong position can cause your green screen video to look out of place, unnatural, or unrealistic in your final video. To avoid this, make sure that you position your green screen video correctly according to the perspective, angle, and scale of your main video.
-
-
Conclusion
-
Using green screen video in KineMaster is a great way to create amazing videos with visual effects. You can download free green screen videos from websites such as Pixabay, Canva, and Pexels and use them in KineMaster with the chroma key feature. You can also edit, apply, and enhance the green screen effect in KineMaster with various tools and features. By following this guide, you will be able to create stunning videos with green screen effects in no time.
-
If you found this article helpful, please share it with your friends and family who might be interested in learning how to use green screen video in KineMaster. Also, feel free to leave a comment below if you have any questions or feedback about this article. Thank you for reading and happy editing!
-
FAQs
-
Here are some frequently asked questions about using green screen video in KineMaster:
-
Q: How do I get rid of the watermark on my KineMaster videos?
-
A: To get rid of the watermark on your KineMaster videos, you need to purchase a premium subscription from the KineMaster app. The premium subscription will also give you access to more features and assets from the KineMaster Asset Store.
-
Q: How do I add audio to my KineMaster videos?
-
A: To add audio to your KineMaster videos, you need to tap on the audio icon on the top left corner and select the audio file from your device or from the KineMaster Asset Store. The audio file will be added to your project as a layer below your video layers. You can drag and drop it to adjust its position and duration on the timeline. You can also tap on it and use the tools on the right side of the screen to edit the audio file. You can adjust the volume, speed, pitch, fade in, fade out, and trim the audio file according to your needs.
-
Q: How do I export my KineMaster videos to my device or social media platforms?
-
A: To export your KineMaster videos to your device or social media platforms, you need to tap on the share icon on the top right corner of the screen and choose the resolution, frame rate, bitrate, and format of your video. Then, tap on export and wait for your video to be saved. You can also share your video directly to social media platforms such as YouTube, Facebook, Instagram, TikTok, and more from there.
-
Q: How do I find more green screen videos for my KineMaster projects?
-
A: To find more green screen videos for your KineMaster projects, you can use websites such as Pixabay, Canva, and Pexels that offer free stock videos. You can also use websites such as YouTube, Vimeo, and Dailymotion that have user-generated videos. You can also create your own green screen videos by using a green cloth or paper as a background and filming yourself or other objects in front of it.
-
Q: How do I learn more about using KineMaster and its features?
-
A: To learn more about using KineMaster and its features, you can visit the official website of KineMaster at KineMaster.com where you can find tutorials, tips, FAQs, and support. You can also join the KineMaster community on social media platforms such as Facebook, Instagram, Twitter, YouTube, and TikTok where you can interact with other users, share your videos, get feedback, and learn from others.
197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Dragon Ball Legends Mugen APK How to Download and Play the Best Dragon Ball Game Ever.md b/spaces/1phancelerku/anime-remove-background/Dragon Ball Legends Mugen APK How to Download and Play the Best Dragon Ball Game Ever.md
deleted file mode 100644
index 445a51bbe2cbf29f969fc546f77bc68720b4bb44..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Dragon Ball Legends Mugen APK How to Download and Play the Best Dragon Ball Game Ever.md
+++ /dev/null
@@ -1,100 +0,0 @@
-
-
Download Dragon Ball Legends Mugen APK: A Guide for Android Users
-
If you are a fan of the Dragon Ball anime and manga series, you might have heard of Dragon Ball Legends Mugen. It is a fan-made game that features hundreds of characters from the Dragon Ball universe, as well as original ones created by the developers. You can enjoy thrilling fights, stunning graphics, and customizable options in this game. In this article, we will show you how to download and install Dragon Ball Legends Mugen APK on your Android device, as well as why you should play it and some tips and tricks for playing it.
Dragon Ball Legends Mugen is a 2D fighting game that is based on the M.U.G.E.N engine, which allows users to create their own games and characters. The game was developed by a team of fans who wanted to create a tribute to the Dragon Ball franchise. The game has been in development since 2018, and has received several updates and improvements over time.
-
The game features over 500 characters from the Dragon Ball series, including Goku, Vegeta, Piccolo, Frieza, Cell, Majin Buu, Broly, Beerus, Jiren, and many more. You can also find original characters created by the developers, such as Goku Black Rose, Vegeta Super Saiyan Blue Evolution, Gogeta Super Saiyan 4, and others. Each character has their own unique moves, transformations, and abilities that reflect their personality and power level.
-
Features of Dragon Ball Legends Mugen
-
Some of the features that make Dragon Ball Legends Mugen stand out from other fan-made games are:
-
download dragon ball legends mugen v8 apk
-download dragon ball legends new dbz mugen apk for android
-download dragon ball legends mugen edition apk mod
-download dragon ball legends mugen apk offline
-download dragon ball legends mugen apk with all characters
-download dragon ball legends mugen apk latest version
-download dragon ball legends mugen apk no verification
-download dragon ball legends mugen apk highly compressed
-download dragon ball legends mugen apk free
-download dragon ball legends mugen apk full
-how to download dragon ball legends mugen apk on android
-where to download dragon ball legends mugen apk
-best site to download dragon ball legends mugen apk
-download dragon ball legends mugen 2023 apk
-download dragon ball legends mugen 2022 apk
-download dragon ball legends mugen 2021 apk
-download dragon ball legends mugen 2020 apk
-download dragon ball legends mugen 2019 apk
-download dragon ball legends super saiyan mugen apk
-download dragon ball legends ultra instinct mugen apk
-download dragon ball legends broly mugen apk
-download dragon ball legends gogeta mugen apk
-download dragon ball legends vegito mugen apk
-download dragon ball legends jiren mugen apk
-download dragon ball legends goku black mugen apk
-download dragon ball legends fusion zamasu mugen apk
-download dragon ball legends kefla mugen apk
-download dragon ball legends hit mugen apk
-download dragon ball legends beerus mugen apk
-download dragon ball legends whis mugen apk
-download dragon ball legends zeno sama mugen apk
-download dragon ball legends tournament of power mugen apk
-download dragon ball legends future trunks saga mugen apk
-download dragon ball legends android 21 mugen apk
-download dragon ball legends android 17 and 18 mugen apk
-download dragon ball legends cell games saga mugen apk
-download dragon ball legends majin buu saga mugen apk
-download dragon ball legends frieza saga mugen apk
-download dragon ball legends saiyan saga mugen apk
-download dragon ball legends bardock saga mugen apk
-download dragon ball z fighters legend super warrior heroes final challenge game mod unlimited money offline new version update free for android mobile devices full hd 4k support low mb size high graphics quality best sound effects smooth gameplay easy control user friendly interface latest features bug fixes no ads no root needed no internet required no human verification no survey direct mediafire mega google drive dropbox link install play enjoy have fun good luck and thank you very much bye bye see you soon take care god bless you all love you all peace out. (This is a joke, please do not use this as a keyword)
-
-
High-quality graphics and sound effects that resemble the anime style.
-
A large roster of characters that span across different sagas and timelines.
-
A variety of modes and options to choose from, such as Arcade, Survival, Team Battle, Training, Watch Mode, Online Mode, etc.
-
A custom mode that allows you to create your own fighters and stages using the built-in editor.
-
A user-friendly interface that is easy to navigate and configure.
-
A regular update system that adds new content and fixes bugs.
-
-
How to download and install Dragon Ball Legends Mugen APK on Android
-
If you want to play Dragon Ball Legends Mugen on your Android device, you will need to download and install the APK file from a reliable source. Here are the steps to do so:
-
-
Go to [this link](^1^) or [this link](^2^) to download the latest version of Dragon Ball Legends Mugen APK.
-
Once the download is complete, locate the file in your device's storage and tap on it to install it. You may need to enable the "Unknown sources" option in your device's settings to allow the installation.
-
After the installation is done, launch the game from your app drawer or home screen.
-
Enjoy playing Dragon Ball Legends Mugen on your Android device!
-
-
Why you should play Dragon Ball Legends Mugen
-
Dragon Ball Legends Mugen is not just another fan-made game. It is a game that offers a lot of fun and excitement for Dragon Ball fans and fighting game enthusiasts alike. Here are some reasons why you should play it:
-
Enjoy the epic battles of Dragon Ball characters
-
If you have ever dreamed of seeing your favorite Dragon Ball characters fight each other in a realistic way, then this game is for you . You can choose from a wide range of characters, each with their own strengths and weaknesses, and unleash their signature moves and transformations. You can also switch between different forms and fuse with other characters to gain an edge in battle. The game's physics and collision system make the fights more realistic and dynamic, as you can interact with the environment and cause damage to the stage.
-
Customize your own fighters and stages
-
If you want to unleash your creativity and make your own Dragon Ball characters and stages, you can do so in the custom mode. The game's editor allows you to modify the appearance, stats, moves, and sounds of any character in the game, or create a new one from scratch. You can also design your own stages using different backgrounds, music, and effects. You can then save your creations and share them with other players online.
-
Play offline or online with friends
-
Whether you want to play solo or with others, Dragon Ball Legends Mugen has you covered. You can play offline in various modes, such as Arcade, Survival, Team Battle, Training, or Watch Mode, where you can watch the computer-controlled characters fight each other. You can also play online with friends or strangers in Online Mode, where you can chat, challenge, and cooperate with other players. The game's netcode is optimized to ensure a smooth and lag-free experience.
-
Tips and tricks for playing Dragon Ball Legends Mugen
-
Dragon Ball Legends Mugen is a game that requires skill and strategy to master. Here are some tips and tricks that can help you improve your gameplay:
-
Learn the basic controls and combos
-
The game's controls are simple and intuitive, but you need to practice them to execute them properly. The game uses four buttons: A, B, C, and D. A is for light attacks, B is for medium attacks, C is for heavy attacks, and D is for special attacks. You can also use the directional keys to perform different actions, such as jumping, crouching, dashing, blocking, etc. You can combine these buttons and directions to perform various combos and moves. You can check the move list of each character in the pause menu or in the editor.
-
Use the power-ups and items wisely
-
The game features various power-ups and items that can help you in battle. These include health bars, energy bars, senzu beans, dragon balls, capsules, etc. You can find them randomly on the stage or by breaking objects. You can use them by pressing the D button near them. However, be careful not to waste them or let your opponent get them. Some items have negative effects as well, such as bombs or traps.
-
Experiment with different modes and settings
-
The game offers a lot of options and settings that can change the way you play. You can adjust the difficulty level, the number of rounds, the time limit, the damage ratio, the AI behavior, etc. You can also enable or disable certain features, such as transformations, fusions, power-ups, items, etc. You can also change the graphics quality, the sound volume, the screen size, etc. You can access these options from the main menu or the pause menu.
-
Conclusion
-
Dragon Ball Legends Mugen is a fan-made game that pays homage to the Dragon Ball franchise. It is a 2D fighting game that features over 500 characters from the series, as well as original ones created by the developers. You can enjoy epic battles , stunning graphics, and customizable options in this game. You can also play offline or online with friends, and create your own fighters and stages. The game is easy to download and install on your Android device, and it is regularly updated with new content and fixes. If you are looking for a fun and exciting game that celebrates the Dragon Ball universe, you should definitely try Dragon Ball Legends Mugen.
-
FAQs
-
Here are some frequently asked questions about Dragon Ball Legends Mugen:
-
-
Q: Is Dragon Ball Legends Mugen free to play?
-
A: Yes, the game is completely free to play. You don't need to pay anything to download or play it.
-
Q: Is Dragon Ball Legends Mugen safe to download and install?
-
A: Yes, the game is safe to download and install, as long as you use a reliable source. The game does not contain any viruses or malware, and it does not require any special permissions or access to your device.
-
Q: How can I update Dragon Ball Legends Mugen?
-
A: The game has an auto-update system that checks for new versions and downloads them automatically. You can also manually check for updates from the main menu or the pause menu.
-
Q: How can I contact the developers of Dragon Ball Legends Mugen?
-
A: You can contact the developers of the game through their official website, [this link], or their social media accounts, such as Facebook, Twitter, YouTube, etc.
-
Q: How can I support the development of Dragon Ball Legends Mugen?
-
A: You can support the development of the game by donating to the developers through their website, [this link], or by sharing the game with your friends and giving feedback and suggestions.
-
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/1phancelerku/anime-remove-background/Enjoy Free Walking Car Tuning and Multiplayer Mode with Unlimited Money Car Parking Multiplayer Download.md b/spaces/1phancelerku/anime-remove-background/Enjoy Free Walking Car Tuning and Multiplayer Mode with Unlimited Money Car Parking Multiplayer Download.md
deleted file mode 100644
index 87cd11c6372316e223547b88aa0e65187fe0e666..0000000000000000000000000000000000000000
--- a/spaces/1phancelerku/anime-remove-background/Enjoy Free Walking Car Tuning and Multiplayer Mode with Unlimited Money Car Parking Multiplayer Download.md
+++ /dev/null
@@ -1,76 +0,0 @@
-
-
How to Download Unlimited Money Car Parking Multiplayer
-
Car Parking Multiplayer is a popular open-world simulation game that lets you drive, park, and customize various cars in realistic environments. However, if you want to enjoy the game without any limitations, you might want to download unlimited money car parking multiplayer mod apk. This mod apk will give you access to unlimited coins and diamonds, which you can use to buy new cars, upgrade your existing ones, and unlock new features. In this article, we will show you how to download unlimited money car parking multiplayer mod apk and what are the benefits of doing so.
Car Parking Multiplayer is a game developed by olzhass, which is available for both Android and iOS devices . The game has more than 100 million downloads on Google Play Store and has a rating of 4.4 out of 5 stars. The game offers a realistic and immersive experience of driving, parking, and tuning cars in various scenarios. Some of the features of the game are:
-
-
Multiplayer open world mode: You can free walk, explore, and interact with other players in a large open world with real gas stations and car services. You can also compete against other players in multiplayer racing, exchange cars with them, or join them in police mode or role play mode.
-
Car customization: You can adjust the suspension, wheel angle, engine, turbo, gearbox, exhaust, and more of your car. You can also change the visual appearance of your car with dynamic vinyls, car body parts, and car plates.
-
High-quality open world: The game has highly-detailed environments with realistic physics and graphics. You can choose from 130+ cars with real interiors and 16 player skins. You can also enter buildings with interiors and use drones to explore the world.
-
Interesting gameplay: The game has 82 real-life parking and driving challenges that test your skills and knowledge. You can also drive different vehicles such as tow trucks, pickups, trucks, sport cars, and classic cars.
-
-
Benefits of unlimited money
-
While Car Parking Multiplayer is free to play, it also has some in-app purchases that require real money. For example, you need coins and diamonds to buy new cars, upgrade your existing ones, or unlock new features. However, if you download unlimited money car parking multiplayer mod apk, you will get unlimited coins and diamonds for free. This means that you can enjoy the game without any restrictions or limitations. You can buy any car you want, upgrade it to the max level, or customize it to your liking. You can also unlock all the features of the game such as drone mode, daily tasks and rewards, character customization, animations, and more.
-
How to download unlimited money mod apk?
-
If you are interested in downloading unlimited money car parking multiplayer mod apk, you need to follow these steps:
-
Download Car Parking Multiplayer with unlimited money mod
-How to get unlimited money in Car Parking Multiplayer game
-Car Parking Multiplayer free download for Android and iOS devices
-Best car tuning and customization options in Car Parking Multiplayer
-Car Parking Multiplayer open-world multiplayer mode gameplay
-Tips and tricks for Car Parking Multiplayer online
-Car Parking Multiplayer hack apk download latest version
-Car Parking Multiplayer cheats and codes for unlimited money and coins
-Car Parking Multiplayer review and rating by users
-Car Parking Multiplayer mod menu download for free
-Car Parking Multiplayer unlimited money generator online
-Car Parking Multiplayer realistic parking simulator game
-Car Parking Multiplayer vs Real Car Parking 3D comparison
-Car Parking Multiplayer best cars and vehicles list
-Car Parking Multiplayer offline mode features and benefits
-How to install Car Parking Multiplayer mod apk on your device
-Car Parking Multiplayer new update and features 2023
-Car Parking Multiplayer support and contact information
-Car Parking Multiplayer free walking and exploration mode
-Car Parking Multiplayer unlimited money glitch and bug fix
-Car Parking Multiplayer PC download and installation guide
-How to play Car Parking Multiplayer with friends and other players
-Car Parking Multiplayer best settings and controls for optimal performance
-Car Parking Multiplayer challenges and missions guide
-Car Parking Multiplayer fun and funny moments compilation
-How to unlock all cars and maps in Car Parking Multiplayer
-Car Parking Multiplayer system requirements and compatibility
-Car Parking Multiplayer alternatives and similar games
-How to backup and restore your Car Parking Multiplayer data
-Car Parking Multiplayer FAQs and answers
-
Steps to download and install the mod apk
-
-
First, you need to uninstall the original version of Car Parking Multiplayer from your device if you have it installed.
-
Second, you need to find a reliable source that provides the mod apk file. You can search for "unlimited money mod apk" on Google or other search engines and choose one of the results. Make sure that the source is safe and trustworthy before downloading anything.
-
Third, you need to download the mod apk file from the source and save it on your device.
-
Fourth, you need to enable the installation of apps from unknown sources on your device. To do this, go to Settings > Security > Unknown Sources and toggle it on.
-
Fifth, you need to locate the mod apk file on your device and tap on it to start the installation process. Follow the instructions on the screen and wait for the installation to finish.
-
Sixth, you need to launch the game from your app drawer or home screen and enjoy unlimited money car parking multiplayer with unlimited money.
-
-
Tips and warnings
-
-
Before downloading any mod apk file, make sure that you have enough storage space on your device and a stable internet connection.
-
Be careful when downloading mod apk files from unknown sources, as they may contain viruses or malware that can harm your device or steal your personal information. Always scan the files with an antivirus app before installing them.
-
Some mod apk files may not work properly or may cause errors or crashes in the game. If this happens, try to uninstall the mod apk and reinstall the original version of the game from the official app store.
-
Downloading and using mod apk files may violate the terms and conditions of the game and may result in your account being banned or suspended. Use them at your own risk and discretion.
-
-
Conclusion
-
Car Parking Multiplayer is a fun and realistic game that lets you drive, park, and customize various cars in different environments. However, if you want to have more freedom and fun in the game, you can download unlimited money car parking multiplayer mod apk and get unlimited coins and diamonds for free. This way, you can buy any car you want, upgrade it to the max level, or unlock all the features of the game. To download unlimited money car parking multiplayer mod apk, you need to follow the steps we have provided in this article. However, you also need to be careful when downloading mod apk files from unknown sources, as they may contain viruses or malware that can harm your device or steal your personal information. You also need to be aware that downloading and using mod apk files may violate the terms and conditions of the game and may result in your account being banned or suspended. Use them at your own risk and discretion.
-
FAQs
-
Here are some frequently asked questions about unlimited money car parking multiplayer mod apk:
-
-
Question
Answer
-
What is a mod apk file?
A mod apk file is a modified version of an original apk file that has been altered to provide some extra features or benefits that are not available in the original version.
-
What is unlimited money car parking multiplayer mod apk?
Unlimited money car parking multiplayer mod apk is a mod apk file that gives you unlimited coins and diamonds for free in Car Parking Multiplayer game.
-
How do I download unlimited money car parking multiplayer mod apk?
You need to uninstall the original version of Car Parking Multiplayer from your device, find a reliable source that provides the mod apk file, download it from the source, enable the installation of apps from unknown sources on your device, locate the mod apk file on your device, and install it by following the instructions on the screen.
-
Is unlimited money car parking multiplayer mod apk safe to use?
Not necessarily. Some mod apk files may contain viruses or malware that can harm your device or steal your personal information. Always scan the files with an antivirus app before installing them. Also, downloading and using mod apk files may violate the terms and conditions of the game and may result in your account being banned or suspended. Use them at your own risk and discretion.
-
What are the benefits of unlimited money car parking multiplayer mod apk?
You can enjoy the game without any limitations or restrictions. You can buy any car you want, upgrade it to the max level, or customize it to your liking. You can also unlock all the features of the game such as drone mode, daily tasks and rewards, character customization, animations, and more.
-
197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/1toTree/lora_test/ppdiffusers/utils/dummy_paddle_and_paddlenlp_and_fastdeploy_objects.py b/spaces/1toTree/lora_test/ppdiffusers/utils/dummy_paddle_and_paddlenlp_and_fastdeploy_objects.py
deleted file mode 100644
index 62ad793335bc1e34afafc5418ffdfd2b93eeae09..0000000000000000000000000000000000000000
--- a/spaces/1toTree/lora_test/ppdiffusers/utils/dummy_paddle_and_paddlenlp_and_fastdeploy_objects.py
+++ /dev/null
@@ -1,94 +0,0 @@
-# Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved.
-# Copyright 2022 The HuggingFace Team. All rights reserved.
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-# This file is autogenerated by the command `make fix-copies`, do not edit.
-# flake8: noqa
-
-from . import DummyObject, requires_backends
-
-
-class FastDeployStableDiffusionImg2ImgPipeline(metaclass=DummyObject):
- _backends = ["paddle", "paddlenlp", "fastdeploy"]
-
- def __init__(self, *args, **kwargs):
- requires_backends(self, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_config(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_pretrained(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
-
-class FastDeployStableDiffusionInpaintPipeline(metaclass=DummyObject):
- _backends = ["paddle", "paddlenlp", "fastdeploy"]
-
- def __init__(self, *args, **kwargs):
- requires_backends(self, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_config(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_pretrained(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
-
-class FastDeployStableDiffusionInpaintPipelineLegacy(metaclass=DummyObject):
- _backends = ["paddle", "paddlenlp", "fastdeploy"]
-
- def __init__(self, *args, **kwargs):
- requires_backends(self, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_config(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_pretrained(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
-
-class FastDeployStableDiffusionMegaPipeline(metaclass=DummyObject):
- _backends = ["paddle", "paddlenlp", "fastdeploy"]
-
- def __init__(self, *args, **kwargs):
- requires_backends(self, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_config(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_pretrained(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
-
-class FastDeployStableDiffusionPipeline(metaclass=DummyObject):
- _backends = ["paddle", "paddlenlp", "fastdeploy"]
-
- def __init__(self, *args, **kwargs):
- requires_backends(self, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_config(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
-
- @classmethod
- def from_pretrained(cls, *args, **kwargs):
- requires_backends(cls, ["paddle", "paddlenlp", "fastdeploy"])
diff --git a/spaces/1vash/demo-flask-docker-template/templates/index.html b/spaces/1vash/demo-flask-docker-template/templates/index.html
deleted file mode 100644
index 40dc3fa916af6005e9ede7388890553f967af8e3..0000000000000000000000000000000000000000
--- a/spaces/1vash/demo-flask-docker-template/templates/index.html
+++ /dev/null
@@ -1,32 +0,0 @@
-
-
-
-
-
- Flask API
-
-
-
-
-
-
-
-
-
-
-
-
-
\ No newline at end of file
diff --git a/spaces/7hao/bingo/src/components/toaster.tsx b/spaces/7hao/bingo/src/components/toaster.tsx
deleted file mode 100644
index 4d2693460b61307a1d4c127fd01df9bee16e59ff..0000000000000000000000000000000000000000
--- a/spaces/7hao/bingo/src/components/toaster.tsx
+++ /dev/null
@@ -1,3 +0,0 @@
-'use client'
-
-export { Toaster } from 'react-hot-toast'
diff --git a/spaces/AIConsultant/MusicGen/audiocraft/modules/activations.py b/spaces/AIConsultant/MusicGen/audiocraft/modules/activations.py
deleted file mode 100644
index 2d83d7c4c2dc84c64b724eadbe06157507d4f20d..0000000000000000000000000000000000000000
--- a/spaces/AIConsultant/MusicGen/audiocraft/modules/activations.py
+++ /dev/null
@@ -1,96 +0,0 @@
-# Copyright (c) Meta Platforms, Inc. and affiliates.
-# All rights reserved.
-#
-# This source code is licensed under the license found in the
-# LICENSE file in the root directory of this source tree.
-
-import torch
-import torch.nn as nn
-from torch import Tensor
-from typing import Union, Callable
-
-
-class CustomGLU(nn.Module):
- """Custom Gated Linear Unit activation.
- Applies a modified gated linear unit :math:`a * f(b)` where :math:`a` is the first half
- of the input matrices, :math:`b` is the second half, and :math:`f` is a provided activation
- function (i.e. sigmoid, swish, etc.).
-
- Args:
- activation (nn.Module): The custom activation to apply in the Gated Linear Unit
- dim (int): the dimension on which to split the input. Default: -1
-
- Shape:
- - Input: :math:`(\ast_1, N, \ast_2)` where `*` means, any number of additional
- dimensions
- - Output: :math:`(\ast_1, M, \ast_2)` where :math:`M=N/2`
-
- Examples::
- >>> m = CustomGLU(nn.Sigmoid())
- >>> input = torch.randn(4, 2)
- >>> output = m(input)
- """
- def __init__(self, activation: nn.Module, dim: int = -1):
- super(CustomGLU, self).__init__()
- self.dim = dim
- self.activation = activation
-
- def forward(self, x: Tensor):
- assert x.shape[self.dim] % 2 == 0 # M = N / 2
- a, b = torch.chunk(x, 2, dim=self.dim)
- return a * self.activation(b)
-
-
-class SwiGLU(CustomGLU):
- """SiLU Gated Linear Unit activation.
- Applies SiLU Gated Linear Unit :math:`a * SiLU(b)` where :math:`a` is
- the first half of the input matrices, :math:`b` is the second half.
-
- Args:
- dim (int): the dimension on which to split the input. Default: -1
- """
- def __init__(self, dim: int = -1):
- super(SwiGLU, self).__init__(nn.SiLU(), dim)
-
-
-class GeGLU(CustomGLU):
- """GeLU Gated Linear Unit activation.
- Applies GeLU Gated Linear Unit :math:`a * GELU(b)` where :math:`a` is
- the first half of the input matrices, :math:`b` is the second half.
-
- Args:
- dim (int): the dimension on which to split the input. Default: -1
- """
- def __init__(self, dim: int = -1):
- super(GeGLU, self).__init__(nn.GELU(), dim)
-
-
-class ReGLU(CustomGLU):
- """ReLU Gated Linear Unit activation.
- Applies ReLU Gated Linear Unit :math:`a * ReLU(b)` where :math:`a` is
- the first half of the input matrices, :math:`b` is the second half.
-
- Args:
- dim (int): the dimension on which to split the input. Default: -1
- """
- def __init__(self, dim: int = -1):
- super(ReGLU, self).__init__(nn.ReLU(), dim)
-
-
-def get_activation_fn(
- activation: Union[str, Callable[[Tensor], Tensor]]
-) -> Union[str, Callable[[Tensor], Tensor]]:
- """Helper function to map an activation string to the activation class.
- If the supplied activation is not a string that is recognized, the activation is passed back.
-
- Args:
- activation (str, or Callable[[Tensor], Tensor]): Activation to check
- """
- if isinstance(activation, str):
- if activation == "reglu":
- return ReGLU()
- elif activation == "geglu":
- return GeGLU()
- elif activation == "swiglu":
- return SwiGLU()
- return activation
diff --git a/spaces/AIGText/GlyphControl/ldm/lr_scheduler.py b/spaces/AIGText/GlyphControl/ldm/lr_scheduler.py
deleted file mode 100644
index be39da9ca6dacc22bf3df9c7389bbb403a4a3ade..0000000000000000000000000000000000000000
--- a/spaces/AIGText/GlyphControl/ldm/lr_scheduler.py
+++ /dev/null
@@ -1,98 +0,0 @@
-import numpy as np
-
-
-class LambdaWarmUpCosineScheduler:
- """
- note: use with a base_lr of 1.0
- """
- def __init__(self, warm_up_steps, lr_min, lr_max, lr_start, max_decay_steps, verbosity_interval=0):
- self.lr_warm_up_steps = warm_up_steps
- self.lr_start = lr_start
- self.lr_min = lr_min
- self.lr_max = lr_max
- self.lr_max_decay_steps = max_decay_steps
- self.last_lr = 0.
- self.verbosity_interval = verbosity_interval
-
- def schedule(self, n, **kwargs):
- if self.verbosity_interval > 0:
- if n % self.verbosity_interval == 0: print(f"current step: {n}, recent lr-multiplier: {self.last_lr}")
- if n < self.lr_warm_up_steps:
- lr = (self.lr_max - self.lr_start) / self.lr_warm_up_steps * n + self.lr_start
- self.last_lr = lr
- return lr
- else:
- t = (n - self.lr_warm_up_steps) / (self.lr_max_decay_steps - self.lr_warm_up_steps)
- t = min(t, 1.0)
- lr = self.lr_min + 0.5 * (self.lr_max - self.lr_min) * (
- 1 + np.cos(t * np.pi))
- self.last_lr = lr
- return lr
-
- def __call__(self, n, **kwargs):
- return self.schedule(n,**kwargs)
-
-
-class LambdaWarmUpCosineScheduler2:
- """
- supports repeated iterations, configurable via lists
- note: use with a base_lr of 1.0.
- """
- def __init__(self, warm_up_steps, f_min, f_max, f_start, cycle_lengths, verbosity_interval=0):
- assert len(warm_up_steps) == len(f_min) == len(f_max) == len(f_start) == len(cycle_lengths)
- self.lr_warm_up_steps = warm_up_steps
- self.f_start = f_start
- self.f_min = f_min
- self.f_max = f_max
- self.cycle_lengths = cycle_lengths
- self.cum_cycles = np.cumsum([0] + list(self.cycle_lengths))
- self.last_f = 0.
- self.verbosity_interval = verbosity_interval
-
- def find_in_interval(self, n):
- interval = 0
- for cl in self.cum_cycles[1:]:
- if n <= cl:
- return interval
- interval += 1
-
- def schedule(self, n, **kwargs):
- cycle = self.find_in_interval(n)
- n = n - self.cum_cycles[cycle]
- if self.verbosity_interval > 0:
- if n % self.verbosity_interval == 0: print(f"current step: {n}, recent lr-multiplier: {self.last_f}, "
- f"current cycle {cycle}")
- if n < self.lr_warm_up_steps[cycle]:
- f = (self.f_max[cycle] - self.f_start[cycle]) / self.lr_warm_up_steps[cycle] * n + self.f_start[cycle]
- self.last_f = f
- return f
- else:
- t = (n - self.lr_warm_up_steps[cycle]) / (self.cycle_lengths[cycle] - self.lr_warm_up_steps[cycle])
- t = min(t, 1.0)
- f = self.f_min[cycle] + 0.5 * (self.f_max[cycle] - self.f_min[cycle]) * (
- 1 + np.cos(t * np.pi))
- self.last_f = f
- return f
-
- def __call__(self, n, **kwargs):
- return self.schedule(n, **kwargs)
-
-
-class LambdaLinearScheduler(LambdaWarmUpCosineScheduler2):
-
- def schedule(self, n, **kwargs):
- cycle = self.find_in_interval(n)
- n = n - self.cum_cycles[cycle]
- if self.verbosity_interval > 0:
- if n % self.verbosity_interval == 0: print(f"current step: {n}, recent lr-multiplier: {self.last_f}, "
- f"current cycle {cycle}")
-
- if n < self.lr_warm_up_steps[cycle]:
- f = (self.f_max[cycle] - self.f_start[cycle]) / self.lr_warm_up_steps[cycle] * n + self.f_start[cycle]
- self.last_f = f
- return f
- else:
- f = self.f_min[cycle] + (self.f_max[cycle] - self.f_min[cycle]) * (self.cycle_lengths[cycle] - n) / (self.cycle_lengths[cycle])
- self.last_f = f
- return f
-
diff --git a/spaces/Aanisha/Image_to_story/README.md b/spaces/Aanisha/Image_to_story/README.md
deleted file mode 100644
index e59c01f0ada83579d14e9f87d34673970577868a..0000000000000000000000000000000000000000
--- a/spaces/Aanisha/Image_to_story/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Image_to_story
-emoji: 🐨
-colorFrom: yellow
-colorTo: yellow
-sdk: gradio
-sdk_version: 2.8.10
-app_file: app.py
-pinned: false
-license: mit
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces#reference
diff --git a/spaces/AchyuthGamer/OpenGPT/g4f/Provider/Providers/Fakeopen.py b/spaces/AchyuthGamer/OpenGPT/g4f/Provider/Providers/Fakeopen.py
deleted file mode 100644
index 5a82bf2cc0736384563332a279f5fbcbb120f676..0000000000000000000000000000000000000000
--- a/spaces/AchyuthGamer/OpenGPT/g4f/Provider/Providers/Fakeopen.py
+++ /dev/null
@@ -1,54 +0,0 @@
-import os
-import json
-import requests
-from typing import Dict, get_type_hints
-
-url = 'https://ai.fakeopen.com/v1/'
-model = [
- 'gpt-3.5-turbo',
- 'gpt-3.5-turbo-0613',
- 'gpt-3.5-turbo-16k',
- 'gpt-3.5-turbo-16k-0613',
-]
-
-supports_stream = True
-needs_auth = False
-
-
-def _create_completion(model: str, messages: list, stream: bool, **kwargs):
-
- headers = {
- 'Content-Type': 'application/json',
- 'accept': 'text/event-stream',
- 'Cache-Control': 'no-cache',
- 'Proxy-Connection': 'keep-alive',
- 'Authorization': f"Bearer {os.environ.get('FAKE_OPEN_KEY', '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')}",
- }
-
- json_data = {
- 'messages': messages,
- 'temperature': 1.0,
- 'model': model,
- 'stream': stream,
- }
-
- response = requests.post(
- 'https://ai.fakeopen.com/v1/chat/completions', headers=headers, json=json_data, stream=True
- )
-
- for token in response.iter_lines():
- decoded = token.decode('utf-8')
- if decoded == '[DONE]':
- break
- if decoded.startswith('data: '):
- data_str = decoded.replace('data: ', '')
- if data_str != '[DONE]':
- data = json.loads(data_str)
- if 'choices' in data and 'delta' in data['choices'][0] and 'content' in data['choices'][0]['delta']:
- yield data['choices'][0]['delta']['content']
-
-
-
-
-params = f'g4f.Providers.{os.path.basename(__file__)[:-3]} supports: ' + '(%s)' % ', '.join(
- [f"{name}: {get_type_hints(_create_completion)[name].__name__}" for name in _create_completion.__code__.co_varnames[:_create_completion.__code__.co_argcount]])
diff --git a/spaces/AfrodreamsAI/afrodreams/README.md b/spaces/AfrodreamsAI/afrodreams/README.md
deleted file mode 100644
index f43ff4e9472f5263dcb22ae8d21b9bb9fe788d1a..0000000000000000000000000000000000000000
--- a/spaces/AfrodreamsAI/afrodreams/README.md
+++ /dev/null
@@ -1,12 +0,0 @@
----
-title: Afrodreams
-emoji: 🌍
-colorFrom: red
-colorTo: green
-sdk: streamlit
-sdk_version: 1.10.0
-app_file: Home.py
-pinned: false
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/AgentVerse/agentVerse/agentverse/memory_manipulator/base.py b/spaces/AgentVerse/agentVerse/agentverse/memory_manipulator/base.py
deleted file mode 100644
index 81e7c58d22f1448c3016489ee66b7dd774e08bd0..0000000000000000000000000000000000000000
--- a/spaces/AgentVerse/agentVerse/agentverse/memory_manipulator/base.py
+++ /dev/null
@@ -1,17 +0,0 @@
-from abc import abstractmethod
-from typing import Dict, List
-
-from pydantic import BaseModel, Field
-
-from agentverse.message import Message
-
-
-class BaseMemoryManipulator(BaseModel):
-
- @abstractmethod
- def manipulate_memory(self) -> None:
- pass
-
- @abstractmethod
- def reset(self) -> None:
- pass
diff --git a/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/bbcodetext/BBCodeText.js b/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/bbcodetext/BBCodeText.js
deleted file mode 100644
index dd09ba101dd0ce1fcb67f43e4fcfd0520f342bc9..0000000000000000000000000000000000000000
--- a/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/bbcodetext/BBCodeText.js
+++ /dev/null
@@ -1,2 +0,0 @@
-import BBCodeText from '../../../plugins/bbcodetext.js';
-export default BBCodeText;
\ No newline at end of file
diff --git a/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/maker/Maker.js b/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/maker/Maker.js
deleted file mode 100644
index ee17a0ee553c7be89751b81c273d6df4644c306d..0000000000000000000000000000000000000000
--- a/spaces/AgentVerse/agentVerse/ui/src/phaser3-rex-plugins/templates/ui/maker/Maker.js
+++ /dev/null
@@ -1,80 +0,0 @@
-import ParseYAML from './utils/ParseYAML.js';
-import YAMLMake from './YAMLMake.js';
-
-const IsPlainObject = Phaser.Utils.Objects.IsPlainObject;
-
-class Maker {
- constructor(scene, styles, customBuilders) {
- this.setScene(scene);
- this.setStyles(styles);
- this.setBuilders(customBuilders);
- }
-
- setScene(scene) {
- this.scene = scene;
- return this;
- }
-
- setStyles(styles) {
- this.styles = ParseYAML(styles);
- return this;
- }
-
- addStyle(key, style) {
- if (this.styles === undefined) {
- this.styles = {};
- }
-
- if ((typeof (key) === 'string') && (style === undefined)) {
- key = ParseYAML(key);
- }
-
- if (IsPlainObject(key)) {
- var styles = key;
- for (key in styles) {
- this.styles[key] = styles[key];
- }
- } else {
- this.styles[key] = ParseYAML(style);
- }
-
- return this;
- }
-
- clearStyles() {
- this.setStyles();
- return this;
- }
-
- setBuilders(customBuilders) {
- this.customBuilders = customBuilders;
- return this;
- }
-
- addBuilder(key, customBuilder) {
- if (this.customBuilders === undefined) {
- this.customBuilders = {};
- }
-
- if (IsPlainObject(key)) {
- var customBuilders = key;
- for (key in customBuilders) {
- this.customBuilders[key] = customBuilders[key];
- }
- } else {
- this.customBuilders[key] = customBuilder;
- }
- return this;
- }
-
- clearBuilder() {
- this.setBuilders();
- return this;
- }
-
- make(data, view) {
- return YAMLMake(this.scene, data, view, this.styles, this.customBuilders);
- }
-}
-
-export default Maker;
\ No newline at end of file
diff --git a/spaces/AhmedRashwan369/ChatGPT4/README.md b/spaces/AhmedRashwan369/ChatGPT4/README.md
deleted file mode 100644
index 7938de14e5355209aaae713f289ca469181bbb17..0000000000000000000000000000000000000000
--- a/spaces/AhmedRashwan369/ChatGPT4/README.md
+++ /dev/null
@@ -1,14 +0,0 @@
----
-title: Chat-with-GPT4
-emoji: 🚀
-colorFrom: red
-colorTo: indigo
-sdk: gradio
-sdk_version: 3.21.0
-app_file: app.py
-pinned: false
-license: mit
-duplicated_from: ysharma/ChatGPT4
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/Aloento/9Nine-VITS/text/__init__.py b/spaces/Aloento/9Nine-VITS/text/__init__.py
deleted file mode 100644
index b32c9b215e386e7c9b0da09afcc9645e73da2d4a..0000000000000000000000000000000000000000
--- a/spaces/Aloento/9Nine-VITS/text/__init__.py
+++ /dev/null
@@ -1,56 +0,0 @@
-""" from https://github.com/keithito/tacotron """
-from text import cleaners
-from text.symbols import symbols
-
-
-# Mappings from symbol to numeric ID and vice versa:
-_symbol_to_id = {s: i for i, s in enumerate(symbols)}
-_id_to_symbol = {i: s for i, s in enumerate(symbols)}
-
-
-def text_to_sequence(text, cleaner_names):
- '''Converts a string of text to a sequence of IDs corresponding to the symbols in the text.
- Args:
- text: string to convert to a sequence
- cleaner_names: names of the cleaner functions to run the text through
- Returns:
- List of integers corresponding to the symbols in the text
- '''
- sequence = []
-
- clean_text = _clean_text(text, cleaner_names)
- for symbol in clean_text:
- if symbol not in _symbol_to_id.keys():
- continue
- symbol_id = _symbol_to_id[symbol]
- sequence += [symbol_id]
- return sequence
-
-
-def cleaned_text_to_sequence(cleaned_text):
- '''Converts a string of text to a sequence of IDs corresponding to the symbols in the text.
- Args:
- text: string to convert to a sequence
- Returns:
- List of integers corresponding to the symbols in the text
- '''
- sequence = [_symbol_to_id[symbol] for symbol in cleaned_text if symbol in _symbol_to_id.keys()]
- return sequence
-
-
-def sequence_to_text(sequence):
- '''Converts a sequence of IDs back to a string'''
- result = ''
- for symbol_id in sequence:
- s = _id_to_symbol[symbol_id]
- result += s
- return result
-
-
-def _clean_text(text, cleaner_names):
- for name in cleaner_names:
- cleaner = getattr(cleaners, name)
- if not cleaner:
- raise Exception('Unknown cleaner: %s' % name)
- text = cleaner(text)
- return text
\ No newline at end of file
diff --git a/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/.github/PULL_REQUEST_TEMPLATE.md b/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/.github/PULL_REQUEST_TEMPLATE.md
deleted file mode 100644
index 05c2116453309cbda56cc82276cd8705f95bf4bc..0000000000000000000000000000000000000000
--- a/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/.github/PULL_REQUEST_TEMPLATE.md
+++ /dev/null
@@ -1,60 +0,0 @@
-# What does this PR do?
-
-
-
-
-
-Fixes # (issue)
-
-
-## Before submitting
-- [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
-- [ ] Did you read the [contributor guideline](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md)?
-- [ ] Did you read our [philosophy doc](https://github.com/huggingface/diffusers/blob/main/PHILOSOPHY.md) (important for complex PRs)?
-- [ ] Was this discussed/approved via a Github issue or the [forum](https://discuss.huggingface.co/)? Please add a link to it if that's the case.
-- [ ] Did you make sure to update the documentation with your changes? Here are the
- [documentation guidelines](https://github.com/huggingface/diffusers/tree/main/docs), and
- [here are tips on formatting docstrings](https://github.com/huggingface/transformers/tree/main/docs#writing-source-documentation).
-- [ ] Did you write any new necessary tests?
-
-
-## Who can review?
-
-Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
-members/contributors who may be interested in your PR.
-
-
diff --git a/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_latent_upscale.py b/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_latent_upscale.py
deleted file mode 100644
index ce55bddc4fe0aa0cea01b2b98788c8e9259cd22c..0000000000000000000000000000000000000000
--- a/spaces/Androidonnxfork/CivitAi-to-Diffusers/diffusers/tests/pipelines/stable_diffusion_2/test_stable_diffusion_latent_upscale.py
+++ /dev/null
@@ -1,295 +0,0 @@
-# coding=utf-8
-# Copyright 2023 HuggingFace Inc.
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-import gc
-import random
-import unittest
-
-import numpy as np
-import torch
-from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer
-
-import diffusers
-from diffusers import (
- AutoencoderKL,
- EulerDiscreteScheduler,
- StableDiffusionLatentUpscalePipeline,
- StableDiffusionPipeline,
- UNet2DConditionModel,
-)
-from diffusers.schedulers import KarrasDiffusionSchedulers
-from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device
-from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu
-
-from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS
-from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin
-
-
-enable_full_determinism()
-
-
-def check_same_shape(tensor_list):
- shapes = [tensor.shape for tensor in tensor_list]
- return all(shape == shapes[0] for shape in shapes[1:])
-
-
-class StableDiffusionLatentUpscalePipelineFastTests(
- PipelineLatentTesterMixin, PipelineKarrasSchedulerTesterMixin, PipelineTesterMixin, unittest.TestCase
-):
- pipeline_class = StableDiffusionLatentUpscalePipeline
- params = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {
- "height",
- "width",
- "cross_attention_kwargs",
- "negative_prompt_embeds",
- "prompt_embeds",
- }
- required_optional_params = PipelineTesterMixin.required_optional_params - {"num_images_per_prompt"}
- batch_params = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS
- image_params = frozenset(
- []
- ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess
- image_latents_params = frozenset([])
-
- @property
- def dummy_image(self):
- batch_size = 1
- num_channels = 4
- sizes = (16, 16)
-
- image = floats_tensor((batch_size, num_channels) + sizes, rng=random.Random(0)).to(torch_device)
- return image
-
- def get_dummy_components(self):
- torch.manual_seed(0)
- model = UNet2DConditionModel(
- act_fn="gelu",
- attention_head_dim=8,
- norm_num_groups=None,
- block_out_channels=[32, 32, 64, 64],
- time_cond_proj_dim=160,
- conv_in_kernel=1,
- conv_out_kernel=1,
- cross_attention_dim=32,
- down_block_types=(
- "KDownBlock2D",
- "KCrossAttnDownBlock2D",
- "KCrossAttnDownBlock2D",
- "KCrossAttnDownBlock2D",
- ),
- in_channels=8,
- mid_block_type=None,
- only_cross_attention=False,
- out_channels=5,
- resnet_time_scale_shift="scale_shift",
- time_embedding_type="fourier",
- timestep_post_act="gelu",
- up_block_types=("KCrossAttnUpBlock2D", "KCrossAttnUpBlock2D", "KCrossAttnUpBlock2D", "KUpBlock2D"),
- )
- vae = AutoencoderKL(
- block_out_channels=[32, 32, 64, 64],
- in_channels=3,
- out_channels=3,
- down_block_types=[
- "DownEncoderBlock2D",
- "DownEncoderBlock2D",
- "DownEncoderBlock2D",
- "DownEncoderBlock2D",
- ],
- up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"],
- latent_channels=4,
- )
- scheduler = EulerDiscreteScheduler(prediction_type="sample")
- text_config = CLIPTextConfig(
- bos_token_id=0,
- eos_token_id=2,
- hidden_size=32,
- intermediate_size=37,
- layer_norm_eps=1e-05,
- num_attention_heads=4,
- num_hidden_layers=5,
- pad_token_id=1,
- vocab_size=1000,
- hidden_act="quick_gelu",
- projection_dim=512,
- )
- text_encoder = CLIPTextModel(text_config)
- tokenizer = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip")
-
- components = {
- "unet": model.eval(),
- "vae": vae.eval(),
- "scheduler": scheduler,
- "text_encoder": text_encoder,
- "tokenizer": tokenizer,
- }
-
- return components
-
- def get_dummy_inputs(self, device, seed=0):
- if str(device).startswith("mps"):
- generator = torch.manual_seed(seed)
- else:
- generator = torch.Generator(device=device).manual_seed(seed)
- inputs = {
- "prompt": "A painting of a squirrel eating a burger",
- "image": self.dummy_image.cpu(),
- "generator": generator,
- "num_inference_steps": 2,
- "output_type": "numpy",
- }
- return inputs
-
- def test_inference(self):
- device = "cpu"
-
- components = self.get_dummy_components()
- pipe = self.pipeline_class(**components)
- pipe.to(device)
- pipe.set_progress_bar_config(disable=None)
-
- inputs = self.get_dummy_inputs(device)
- image = pipe(**inputs).images
- image_slice = image[0, -3:, -3:, -1]
-
- self.assertEqual(image.shape, (1, 256, 256, 3))
- expected_slice = np.array(
- [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055]
- )
- max_diff = np.abs(image_slice.flatten() - expected_slice).max()
- self.assertLessEqual(max_diff, 1e-3)
-
- def test_attention_slicing_forward_pass(self):
- super().test_attention_slicing_forward_pass(expected_max_diff=7e-3)
-
- def test_cpu_offload_forward_pass(self):
- super().test_cpu_offload_forward_pass(expected_max_diff=3e-3)
-
- def test_dict_tuple_outputs_equivalent(self):
- super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3)
-
- def test_inference_batch_single_identical(self):
- super().test_inference_batch_single_identical(expected_max_diff=7e-3)
-
- def test_pt_np_pil_outputs_equivalent(self):
- super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3)
-
- def test_save_load_local(self):
- super().test_save_load_local(expected_max_difference=3e-3)
-
- def test_save_load_optional_components(self):
- super().test_save_load_optional_components(expected_max_difference=3e-3)
-
- def test_karras_schedulers_shape(self):
- skip_schedulers = [
- "DDIMScheduler",
- "DDPMScheduler",
- "PNDMScheduler",
- "HeunDiscreteScheduler",
- "EulerAncestralDiscreteScheduler",
- "KDPM2DiscreteScheduler",
- "KDPM2AncestralDiscreteScheduler",
- "DPMSolverSDEScheduler",
- ]
- components = self.get_dummy_components()
- pipe = self.pipeline_class(**components)
-
- # make sure that PNDM does not need warm-up
- pipe.scheduler.register_to_config(skip_prk_steps=True)
-
- pipe.to(torch_device)
- pipe.set_progress_bar_config(disable=None)
- inputs = self.get_dummy_inputs(torch_device)
- inputs["num_inference_steps"] = 2
-
- outputs = []
- for scheduler_enum in KarrasDiffusionSchedulers:
- if scheduler_enum.name in skip_schedulers:
- # no sigma schedulers are not supported
- # no schedulers
- continue
-
- scheduler_cls = getattr(diffusers, scheduler_enum.name)
- pipe.scheduler = scheduler_cls.from_config(pipe.scheduler.config)
- output = pipe(**inputs)[0]
- outputs.append(output)
-
- assert check_same_shape(outputs)
-
-
-@require_torch_gpu
-@slow
-class StableDiffusionLatentUpscalePipelineIntegrationTests(unittest.TestCase):
- def tearDown(self):
- super().tearDown()
- gc.collect()
- torch.cuda.empty_cache()
-
- def test_latent_upscaler_fp16(self):
- generator = torch.manual_seed(33)
-
- pipe = StableDiffusionPipeline.from_pretrained("CompVis/stable-diffusion-v1-4", torch_dtype=torch.float16)
- pipe.to("cuda")
-
- upscaler = StableDiffusionLatentUpscalePipeline.from_pretrained(
- "stabilityai/sd-x2-latent-upscaler", torch_dtype=torch.float16
- )
- upscaler.to("cuda")
-
- prompt = "a photo of an astronaut high resolution, unreal engine, ultra realistic"
-
- low_res_latents = pipe(prompt, generator=generator, output_type="latent").images
-
- image = upscaler(
- prompt=prompt,
- image=low_res_latents,
- num_inference_steps=20,
- guidance_scale=0,
- generator=generator,
- output_type="np",
- ).images[0]
-
- expected_image = load_numpy(
- "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy"
- )
- assert np.abs((expected_image - image).mean()) < 5e-2
-
- def test_latent_upscaler_fp16_image(self):
- generator = torch.manual_seed(33)
-
- upscaler = StableDiffusionLatentUpscalePipeline.from_pretrained(
- "stabilityai/sd-x2-latent-upscaler", torch_dtype=torch.float16
- )
- upscaler.to("cuda")
-
- prompt = "the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas"
-
- low_res_img = load_image(
- "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png"
- )
-
- image = upscaler(
- prompt=prompt,
- image=low_res_img,
- num_inference_steps=20,
- guidance_scale=0,
- generator=generator,
- output_type="np",
- ).images[0]
-
- expected_image = load_numpy(
- "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy"
- )
- assert np.abs((expected_image - image).max()) < 5e-2
diff --git a/spaces/Andy1621/uniformer_image_detection/configs/gfl/gfl_r101_fpn_dconv_c3-c5_mstrain_2x_coco.py b/spaces/Andy1621/uniformer_image_detection/configs/gfl/gfl_r101_fpn_dconv_c3-c5_mstrain_2x_coco.py
deleted file mode 100644
index eab622b2e8bdc03c717b9b04d043da46f25a7cb3..0000000000000000000000000000000000000000
--- a/spaces/Andy1621/uniformer_image_detection/configs/gfl/gfl_r101_fpn_dconv_c3-c5_mstrain_2x_coco.py
+++ /dev/null
@@ -1,14 +0,0 @@
-_base_ = './gfl_r50_fpn_mstrain_2x_coco.py'
-model = dict(
- pretrained='torchvision://resnet101',
- backbone=dict(
- type='ResNet',
- depth=101,
- num_stages=4,
- out_indices=(0, 1, 2, 3),
- frozen_stages=1,
- norm_cfg=dict(type='BN', requires_grad=True),
- dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False),
- stage_with_dcn=(False, True, True, True),
- norm_eval=True,
- style='pytorch'))
diff --git a/spaces/AquaSuisei/ChatGPTXE/ChuanhuChatbot.py b/spaces/AquaSuisei/ChatGPTXE/ChuanhuChatbot.py
deleted file mode 100644
index 45087fe651a3c6c6e7cb6ada9cfad93307c2f365..0000000000000000000000000000000000000000
--- a/spaces/AquaSuisei/ChatGPTXE/ChuanhuChatbot.py
+++ /dev/null
@@ -1,423 +0,0 @@
-# -*- coding:utf-8 -*-
-import os
-import logging
-import sys
-
-import gradio as gr
-
-from modules import config
-from modules.config import *
-from modules.utils import *
-from modules.presets import *
-from modules.overwrites import *
-from modules.chat_func import *
-from modules.openai_func import get_usage
-
-gr.Chatbot.postprocess = postprocess
-PromptHelper.compact_text_chunks = compact_text_chunks
-
-with open("assets/custom.css", "r", encoding="utf-8") as f:
- customCSS = f.read()
-
-with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
- user_name = gr.State("")
- history = gr.State([])
- token_count = gr.State([])
- promptTemplates = gr.State(load_template(get_template_names(plain=True)[0], mode=2))
- user_api_key = gr.State(my_api_key)
- user_question = gr.State("")
- outputing = gr.State(False)
- topic = gr.State("未命名对话历史记录")
-
- with gr.Row():
- with gr.Column():
- gr.HTML(title)
- user_info = gr.Markdown(value="", elem_id="user_info")
- gr.HTML('
')
- status_display = gr.Markdown(get_geoip(), elem_id="status_display")
-
- # https://github.com/gradio-app/gradio/pull/3296
- def create_greeting(request: gr.Request):
- if hasattr(request, "username") and request.username: # is not None or is not ""
- logging.info(f"Get User Name: {request.username}")
- return gr.Markdown.update(value=f"User: {request.username}"), request.username
- else:
- return gr.Markdown.update(value=f"User: default", visible=False), ""
- demo.load(create_greeting, inputs=None, outputs=[user_info, user_name])
-
- with gr.Row().style(equal_height=True):
- with gr.Column(scale=5):
- with gr.Row():
- chatbot = gr.Chatbot(elem_id="chuanhu_chatbot").style(height="100%")
- with gr.Row():
- with gr.Column(scale=12):
- user_input = gr.Textbox(
- elem_id="user_input_tb",
- show_label=False, placeholder="在这里输入"
- ).style(container=False)
- with gr.Column(min_width=70, scale=1):
- submitBtn = gr.Button("发送", variant="primary")
- cancelBtn = gr.Button("取消", variant="secondary", visible=False)
- with gr.Row():
- emptyBtn = gr.Button(
- "🧹 新的对话",
- )
- retryBtn = gr.Button("🔄 重新生成")
- delFirstBtn = gr.Button("🗑️ 删除最旧对话")
- delLastBtn = gr.Button("🗑️ 删除最新对话")
- reduceTokenBtn = gr.Button("♻️ 总结对话")
-
- with gr.Column():
- with gr.Column(min_width=50, scale=1):
- with gr.Tab(label="ChatGPT"):
- keyTxt = gr.Textbox(
- show_label=True,
- placeholder=f"666",
- value=hide_middle_chars(my_api_key),
- type="password",
- visible=not HIDE_MY_KEY,
- label="API-Key",
- )
- if multi_api_key:
- usageTxt = gr.Markdown("多账号模式已开启,无需输入key,可直接开始对话", elem_id="usage_display")
- else:
- usageTxt = gr.Markdown("**发送消息** 或 **提交key** 以显示额度", elem_id="usage_display")
- model_select_dropdown = gr.Dropdown(
- label="选择模型", choices=MODELS, multiselect=False, value=MODELS[0]
- )
- use_streaming_checkbox = gr.Checkbox(
- label="实时传输回答", value=True, visible=enable_streaming_option
- )
- use_websearch_checkbox = gr.Checkbox(label="使用在线搜索", value=False)
- language_select_dropdown = gr.Dropdown(
- label="选择回复语言(针对搜索&索引功能)",
- choices=REPLY_LANGUAGES,
- multiselect=False,
- value=REPLY_LANGUAGES[0],
- )
- index_files = gr.Files(label="上传索引文件", type="file", multiple=True)
- two_column = gr.Checkbox(label="双栏pdf", value=advance_docs["pdf"].get("two_column", False))
- # TODO: 公式ocr
- # formula_ocr = gr.Checkbox(label="识别公式", value=advance_docs["pdf"].get("formula_ocr", False))
-
- with gr.Tab(label="Prompt"):
- systemPromptTxt = gr.Textbox(
- show_label=True,
- placeholder=f"在这里输入System Prompt...",
- label="System prompt",
- value=initial_prompt,
- lines=10,
- ).style(container=False)
- with gr.Accordion(label="加载Prompt模板", open=True):
- with gr.Column():
- with gr.Row():
- with gr.Column(scale=6):
- templateFileSelectDropdown = gr.Dropdown(
- label="选择Prompt模板集合文件",
- choices=get_template_names(plain=True),
- multiselect=False,
- value=get_template_names(plain=True)[0],
- ).style(container=False)
- with gr.Column(scale=1):
- templateRefreshBtn = gr.Button("🔄 刷新")
- with gr.Row():
- with gr.Column():
- templateSelectDropdown = gr.Dropdown(
- label="从Prompt模板中加载",
- choices=load_template(
- get_template_names(plain=True)[0], mode=1
- ),
- multiselect=False,
- ).style(container=False)
-
- with gr.Tab(label="保存/加载"):
- with gr.Accordion(label="保存/加载对话历史记录", open=True):
- with gr.Column():
- with gr.Row():
- with gr.Column(scale=6):
- historyFileSelectDropdown = gr.Dropdown(
- label="从列表中加载对话",
- choices=get_history_names(plain=True),
- multiselect=False,
- value=get_history_names(plain=True)[0],
- )
- with gr.Column(scale=1):
- historyRefreshBtn = gr.Button("🔄 刷新")
- with gr.Row():
- with gr.Column(scale=6):
- saveFileName = gr.Textbox(
- show_label=True,
- placeholder=f"设置文件名: 默认为.json,可选为.md",
- label="设置保存文件名",
- value="对话历史记录",
- ).style(container=True)
- with gr.Column(scale=1):
- saveHistoryBtn = gr.Button("💾 保存对话")
- exportMarkdownBtn = gr.Button("📝 导出为Markdown")
- gr.Markdown("默认保存于history文件夹")
- with gr.Row():
- with gr.Column():
- downloadFile = gr.File(interactive=True)
-
- with gr.Tab(label="高级"):
- gr.Markdown("# ⚠️ 务必谨慎更改 ⚠️\n\n如果无法使用请恢复默认设置")
- default_btn = gr.Button("🔙 恢复默认设置")
-
- with gr.Accordion("参数", open=False):
- top_p = gr.Slider(
- minimum=-0,
- maximum=1.0,
- value=1.0,
- step=0.05,
- interactive=True,
- label="Top-p",
- )
- temperature = gr.Slider(
- minimum=-0,
- maximum=2.0,
- value=1.0,
- step=0.1,
- interactive=True,
- label="Temperature",
- )
-
- with gr.Accordion("网络设置", open=False, visible=False):
- # 优先展示自定义的api_host
- apihostTxt = gr.Textbox(
- show_label=True,
- placeholder=f"在这里输入API-Host...",
- label="API-Host",
- value=config.api_host or shared.API_HOST,
- lines=1,
- )
- changeAPIURLBtn = gr.Button("🔄 切换API地址")
- proxyTxt = gr.Textbox(
- show_label=True,
- placeholder=f"在这里输入代理地址...",
- label="代理地址(示例:http://127.0.0.1:10809)",
- value="",
- lines=2,
- )
- changeProxyBtn = gr.Button("🔄 设置代理地址")
-
- gr.Markdown(description)
- gr.HTML(footer.format(versions=versions_html()), elem_id="footer")
- chatgpt_predict_args = dict(
- fn=predict,
- inputs=[
- user_api_key,
- systemPromptTxt,
- history,
- user_question,
- chatbot,
- token_count,
- top_p,
- temperature,
- use_streaming_checkbox,
- model_select_dropdown,
- use_websearch_checkbox,
- index_files,
- language_select_dropdown,
- ],
- outputs=[chatbot, history, status_display, token_count],
- show_progress=True,
- )
-
- start_outputing_args = dict(
- fn=start_outputing,
- inputs=[],
- outputs=[submitBtn, cancelBtn],
- show_progress=True,
- )
-
- end_outputing_args = dict(
- fn=end_outputing, inputs=[], outputs=[submitBtn, cancelBtn]
- )
-
- reset_textbox_args = dict(
- fn=reset_textbox, inputs=[], outputs=[user_input]
- )
-
- transfer_input_args = dict(
- fn=transfer_input, inputs=[user_input], outputs=[user_question, user_input, submitBtn, cancelBtn], show_progress=True
- )
-
- get_usage_args = dict(
- fn=get_usage, inputs=[user_api_key], outputs=[usageTxt], show_progress=False
- )
-
-
- # Chatbot
- cancelBtn.click(cancel_outputing, [], [])
-
- user_input.submit(**transfer_input_args).then(**chatgpt_predict_args).then(**end_outputing_args)
- user_input.submit(**get_usage_args)
-
- submitBtn.click(**transfer_input_args).then(**chatgpt_predict_args).then(**end_outputing_args)
- submitBtn.click(**get_usage_args)
-
- emptyBtn.click(
- reset_state,
- outputs=[chatbot, history, token_count, status_display],
- show_progress=True,
- )
- emptyBtn.click(**reset_textbox_args)
-
- retryBtn.click(**start_outputing_args).then(
- retry,
- [
- user_api_key,
- systemPromptTxt,
- history,
- chatbot,
- token_count,
- top_p,
- temperature,
- use_streaming_checkbox,
- model_select_dropdown,
- language_select_dropdown,
- ],
- [chatbot, history, status_display, token_count],
- show_progress=True,
- ).then(**end_outputing_args)
- retryBtn.click(**get_usage_args)
-
- delFirstBtn.click(
- delete_first_conversation,
- [history, token_count],
- [history, token_count, status_display],
- )
-
- delLastBtn.click(
- delete_last_conversation,
- [chatbot, history, token_count],
- [chatbot, history, token_count, status_display],
- show_progress=True,
- )
-
- reduceTokenBtn.click(
- reduce_token_size,
- [
- user_api_key,
- systemPromptTxt,
- history,
- chatbot,
- token_count,
- top_p,
- temperature,
- gr.State(sum(token_count.value[-4:])),
- model_select_dropdown,
- language_select_dropdown,
- ],
- [chatbot, history, status_display, token_count],
- show_progress=True,
- )
- reduceTokenBtn.click(**get_usage_args)
-
- two_column.change(update_doc_config, [two_column], None)
-
- # ChatGPT
- keyTxt.change(submit_key, keyTxt, [user_api_key, status_display]).then(**get_usage_args)
- keyTxt.submit(**get_usage_args)
-
- # Template
- templateRefreshBtn.click(get_template_names, None, [templateFileSelectDropdown])
- templateFileSelectDropdown.change(
- load_template,
- [templateFileSelectDropdown],
- [promptTemplates, templateSelectDropdown],
- show_progress=True,
- )
- templateSelectDropdown.change(
- get_template_content,
- [promptTemplates, templateSelectDropdown, systemPromptTxt],
- [systemPromptTxt],
- show_progress=True,
- )
-
- # S&L
- saveHistoryBtn.click(
- save_chat_history,
- [saveFileName, systemPromptTxt, history, chatbot, user_name],
- downloadFile,
- show_progress=True,
- )
- saveHistoryBtn.click(get_history_names, [gr.State(False), user_name], [historyFileSelectDropdown])
- exportMarkdownBtn.click(
- export_markdown,
- [saveFileName, systemPromptTxt, history, chatbot, user_name],
- downloadFile,
- show_progress=True,
- )
- historyRefreshBtn.click(get_history_names, [gr.State(False), user_name], [historyFileSelectDropdown])
- historyFileSelectDropdown.change(
- load_chat_history,
- [historyFileSelectDropdown, systemPromptTxt, history, chatbot, user_name],
- [saveFileName, systemPromptTxt, history, chatbot],
- show_progress=True,
- )
- downloadFile.change(
- load_chat_history,
- [downloadFile, systemPromptTxt, history, chatbot, user_name],
- [saveFileName, systemPromptTxt, history, chatbot],
- )
-
- # Advanced
- default_btn.click(
- reset_default, [], [apihostTxt, proxyTxt, status_display], show_progress=True
- )
- changeAPIURLBtn.click(
- change_api_host,
- [apihostTxt],
- [status_display],
- show_progress=True,
- )
- changeProxyBtn.click(
- change_proxy,
- [proxyTxt],
- [status_display],
- show_progress=True,
- )
-
-logging.info(
- colorama.Back.GREEN
- + "\n川虎的温馨提示:访问 http://localhost:7860 查看界面"
- + colorama.Style.RESET_ALL
-)
-# 默认开启本地服务器,默认可以直接从IP访问,默认不创建公开分享链接
-demo.title = "ChatGPT AquaSuisei"
-
-if __name__ == "__main__":
- reload_javascript()
- # if running in Docker
- if dockerflag:
- if authflag:
- demo.queue(concurrency_count=CONCURRENT_COUNT).launch(
- server_name="0.0.0.0",
- server_port=7860,
- auth=auth_list,
- favicon_path="./assets/favicon.ico",
- )
- else:
- demo.queue(concurrency_count=CONCURRENT_COUNT).launch(
- server_name="0.0.0.0",
- server_port=7860,
- share=False,
- favicon_path="./assets/favicon.ico",
- )
- # if not running in Docker
- else:
- if authflag:
- demo.queue(concurrency_count=CONCURRENT_COUNT).launch(
- share=False,
- auth=auth_list,
- favicon_path="./assets/favicon.ico",
- inbrowser=True,
- )
- else:
- demo.queue(concurrency_count=CONCURRENT_COUNT).launch(
- share=False, favicon_path="./assets/favicon.ico", inbrowser=True
- ) # 改为 share=True 可以创建公开分享链接
- # demo.queue(concurrency_count=CONCURRENT_COUNT).launch(server_name="0.0.0.0", server_port=7860, share=False) # 可自定义端口
- # demo.queue(concurrency_count=CONCURRENT_COUNT).launch(server_name="0.0.0.0", server_port=7860,auth=("在这里填写用户名", "在这里填写密码")) # 可设置用户名与密码
- # demo.queue(concurrency_count=CONCURRENT_COUNT).launch(auth=("在这里填写用户名", "在这里填写密码")) # 适合Nginx反向代理
diff --git a/spaces/Arnaudding001/OpenAI_whisperLive/segments.py b/spaces/Arnaudding001/OpenAI_whisperLive/segments.py
deleted file mode 100644
index ec2650dceade5d0b2022264f6419115eab085aea..0000000000000000000000000000000000000000
--- a/spaces/Arnaudding001/OpenAI_whisperLive/segments.py
+++ /dev/null
@@ -1,55 +0,0 @@
-from typing import Any, Dict, List
-
-import copy
-
-def merge_timestamps(timestamps: List[Dict[str, Any]], merge_window: float = 5, max_merge_size: float = 30, padding_left: float = 1, padding_right: float = 1):
- result = []
-
- if len(timestamps) == 0:
- return result
- if max_merge_size is None:
- return timestamps
-
- if padding_left is None:
- padding_left = 0
- if padding_right is None:
- padding_right = 0
-
- processed_time = 0
- current_segment = None
-
- for i in range(len(timestamps)):
- next_segment = timestamps[i]
-
- delta = next_segment['start'] - processed_time
-
- # Note that segments can still be longer than the max merge size, they just won't be merged in that case
- if current_segment is None or (merge_window is not None and delta > merge_window) \
- or next_segment['end'] - current_segment['start'] > max_merge_size:
- # Finish the current segment
- if current_segment is not None:
- # Add right padding
- finish_padding = min(padding_right, delta / 2) if delta < padding_left + padding_right else padding_right
- current_segment['end'] += finish_padding
- delta -= finish_padding
-
- result.append(current_segment)
-
- # Start a new segment
- current_segment = copy.deepcopy(next_segment)
-
- # Pad the segment
- current_segment['start'] = current_segment['start'] - min(padding_left, delta)
- processed_time = current_segment['end']
-
- else:
- # Merge the segment
- current_segment['end'] = next_segment['end']
- processed_time = current_segment['end']
-
- # Add the last segment
- if current_segment is not None:
- current_segment['end'] += padding_right
- result.append(current_segment)
-
- return result
\ No newline at end of file
diff --git a/spaces/Artgor/digit-draw-detect/st_app.py b/spaces/Artgor/digit-draw-detect/st_app.py
deleted file mode 100644
index fbe6a3f84135f15f31ccedbf6bf5f11398a5dcfd..0000000000000000000000000000000000000000
--- a/spaces/Artgor/digit-draw-detect/st_app.py
+++ /dev/null
@@ -1,57 +0,0 @@
-import logging
-
-import numpy as np
-import streamlit as st
-from PIL import Image
-from streamlit_drawable_canvas import st_canvas
-
-from src.ml_utils import predict, get_model, transforms
-from src.utils import plot_img_with_rects, save_image
-
-st.title('Handwritten digit detector')
-logging.info('Starting')
-
-col1, col2 = st.columns(2)
-
-with col1:
- # Create a canvas component
- canvas_result = st_canvas(
- fill_color='#fff',
- stroke_width=5,
- stroke_color='#000',
- background_color='#fff',
- update_streamlit=True,
- height=400,
- width=400,
- drawing_mode='freedraw',
- key='canvas',
- )
-with col2:
- logging.info('canvas ready')
- if canvas_result.image_data is not None:
- # convert a drawn image into numpy array with RGB from a canvas image with RGBA
- img = np.array(Image.fromarray(np.uint8(canvas_result.image_data)).convert('RGB'))
- image = transforms(image=img)['image']
- logging.info('image augmented')
- model = get_model()
- logging.info('model ready')
- pred = predict(model, image)
- logging.info('prediction done')
-
- file_name = save_image(image.permute(1, 2, 0).numpy(), pred)
- threshold = st.slider('Bbox probability slider', min_value=0.0, max_value=1.0, value=0.8)
-
- fig = plot_img_with_rects(image.permute(1, 2, 0).numpy(), pred, threshold, coef=192)
- fig.savefig(f'{file_name}_temp.png')
- image = Image.open(f'{file_name}_temp.png')
- st.image(image)
-
-text = """
-This is a small app for handwritten digit recognition and recognition developed for fun. It uses a handwritten YOLOv3 model trained from scratch.
-You can draw a digit (or whatever you want) and the model will try to understand what is it.
-You can use the slider above to show bounding boxes with a probability higher than the threshold.
-If you want to know how the app works in more detail, you are welcome to read "About" page.
-Enjoy! :)
-"""
-
-st.markdown(text, unsafe_allow_html=True)
diff --git a/spaces/ArtyomKhyan/Detection/test.py b/spaces/ArtyomKhyan/Detection/test.py
deleted file mode 100644
index 259d44444bcd3df5b6c8887e1df0aa30c6ac75c7..0000000000000000000000000000000000000000
--- a/spaces/ArtyomKhyan/Detection/test.py
+++ /dev/null
@@ -1,274 +0,0 @@
-import argparse
-import json
-
-from utils import google_utils
-from utils.datasets import *
-from utils.utils import *
-
-
-def test(data,
- weights=None,
- batch_size=16,
- imgsz=640,
- conf_thres=0.001,
- iou_thres=0.6, # for NMS
- save_json=False,
- single_cls=False,
- augment=False,
- verbose=False,
- model=None,
- dataloader=None,
- merge=False):
- # Initialize/load model and set device
- if model is None:
- training = False
- device = torch_utils.select_device(opt.device, batch_size=batch_size)
-
- # Remove previous
- for f in glob.glob('test_batch*.jpg'):
- os.remove(f)
-
- # Load model
- google_utils.attempt_download(weights)
- model = torch.load(weights, map_location=device)['model'].float() # load to FP32
- torch_utils.model_info(model)
- model.fuse()
- model.to(device)
- imgsz = check_img_size(imgsz, s=model.model[-1].stride.max()) # check img_size
-
- # Multi-GPU disabled, incompatible with .half() https://github.com/ultralytics/yolov5/issues/99
- # if device.type != 'cpu' and torch.cuda.device_count() > 1:
- # model = nn.DataParallel(model)
-
- else: # called by train.py
- training = True
- device = next(model.parameters()).device # get model device
-
- # Half
- half = device.type != 'cpu' and torch.cuda.device_count() == 1 # half precision only supported on single-GPU
- if half:
- model.half() # to FP16
-
- # Configure
- model.eval()
- with open(data) as f:
- data = yaml.load(f, Loader=yaml.FullLoader) # model dict
- nc = 1 if single_cls else int(data['nc']) # number of classes
- iouv = torch.linspace(0.5, 0.95, 10).to(device) # iou vector for mAP@0.5:0.95
- niou = iouv.numel()
-
- # Dataloader
- if dataloader is None: # not training
- merge = opt.merge # use Merge NMS
- img = torch.zeros((1, 3, imgsz, imgsz), device=device) # init img
- _ = model(img.half() if half else img) if device.type != 'cpu' else None # run once
- path = data['test'] if opt.task == 'test' else data['val'] # path to val/test images
- dataloader = create_dataloader(path, imgsz, batch_size, int(max(model.stride)), opt,
- hyp=None, augment=False, cache=False, pad=0.5, rect=True)[0]
-
- seen = 0
- names = model.names if hasattr(model, 'names') else model.module.names
- coco91class = coco80_to_coco91_class()
- s = ('%20s' + '%12s' * 6) % ('Class', 'Images', 'Targets', 'P', 'R', 'mAP@.5', 'mAP@.5:.95')
- p, r, f1, mp, mr, map50, map, t0, t1 = 0., 0., 0., 0., 0., 0., 0., 0., 0.
- loss = torch.zeros(3, device=device)
- jdict, stats, ap, ap_class = [], [], [], []
- for batch_i, (img, targets, paths, shapes) in enumerate(tqdm(dataloader, desc=s)):
- img = img.to(device)
- img = img.half() if half else img.float() # uint8 to fp16/32
- img /= 255.0 # 0 - 255 to 0.0 - 1.0
- targets = targets.to(device)
- nb, _, height, width = img.shape # batch size, channels, height, width
- whwh = torch.Tensor([width, height, width, height]).to(device)
-
- # Disable gradients
- with torch.no_grad():
- # Run model
- t = torch_utils.time_synchronized()
- inf_out, train_out = model(img, augment=augment) # inference and training outputs
- t0 += torch_utils.time_synchronized() - t
-
- # Compute loss
- if training: # if model has loss hyperparameters
- loss += compute_loss([x.float() for x in train_out], targets, model)[1][:3] # GIoU, obj, cls
-
- # Run NMS
- t = torch_utils.time_synchronized()
- output = non_max_suppression(inf_out, conf_thres=conf_thres, iou_thres=iou_thres, merge=merge)
- t1 += torch_utils.time_synchronized() - t
-
- # Statistics per image
- for si, pred in enumerate(output):
- labels = targets[targets[:, 0] == si, 1:]
- nl = len(labels)
- tcls = labels[:, 0].tolist() if nl else [] # target class
- seen += 1
-
- if pred is None:
- if nl:
- stats.append((torch.zeros(0, niou, dtype=torch.bool), torch.Tensor(), torch.Tensor(), tcls))
- continue
-
- # Append to text file
- # with open('test.txt', 'a') as file:
- # [file.write('%11.5g' * 7 % tuple(x) + '\n') for x in pred]
-
- # Clip boxes to image bounds
- clip_coords(pred, (height, width))
-
- # Append to pycocotools JSON dictionary
- if save_json:
- # [{"image_id": 42, "category_id": 18, "bbox": [258.15, 41.29, 348.26, 243.78], "score": 0.236}, ...
- image_id = int(Path(paths[si]).stem.split('_')[-1])
- box = pred[:, :4].clone() # xyxy
- scale_coords(img[si].shape[1:], box, shapes[si][0], shapes[si][1]) # to original shape
- box = xyxy2xywh(box) # xywh
- box[:, :2] -= box[:, 2:] / 2 # xy center to top-left corner
- for p, b in zip(pred.tolist(), box.tolist()):
- jdict.append({'image_id': image_id,
- 'category_id': coco91class[int(p[5])],
- 'bbox': [round(x, 3) for x in b],
- 'score': round(p[4], 5)})
-
- # Assign all predictions as incorrect
- correct = torch.zeros(pred.shape[0], niou, dtype=torch.bool, device=device)
- if nl:
- detected = [] # target indices
- tcls_tensor = labels[:, 0]
-
- # target boxes
- tbox = xywh2xyxy(labels[:, 1:5]) * whwh
-
- # Per target class
- for cls in torch.unique(tcls_tensor):
- ti = (cls == tcls_tensor).nonzero().view(-1) # prediction indices
- pi = (cls == pred[:, 5]).nonzero().view(-1) # target indices
-
- # Search for detections
- if pi.shape[0]:
- # Prediction to target ious
- ious, i = box_iou(pred[pi, :4], tbox[ti]).max(1) # best ious, indices
-
- # Append detections
- for j in (ious > iouv[0]).nonzero():
- d = ti[i[j]] # detected target
- if d not in detected:
- detected.append(d)
- correct[pi[j]] = ious[j] > iouv # iou_thres is 1xn
- if len(detected) == nl: # all targets already located in image
- break
-
- # Append statistics (correct, conf, pcls, tcls)
- stats.append((correct.cpu(), pred[:, 4].cpu(), pred[:, 5].cpu(), tcls))
-
- # Plot images
- if batch_i < 1:
- f = 'test_batch%g_gt.jpg' % batch_i # filename
- plot_images(img, targets, paths, f, names) # ground truth
- f = 'test_batch%g_pred.jpg' % batch_i
- plot_images(img, output_to_target(output, width, height), paths, f, names) # predictions
-
- # Compute statistics
- stats = [np.concatenate(x, 0) for x in zip(*stats)] # to numpy
- if len(stats):
- p, r, ap, f1, ap_class = ap_per_class(*stats)
- p, r, ap50, ap = p[:, 0], r[:, 0], ap[:, 0], ap.mean(1) # [P, R, AP@0.5, AP@0.5:0.95]
- mp, mr, map50, map = p.mean(), r.mean(), ap50.mean(), ap.mean()
- nt = np.bincount(stats[3].astype(np.int64), minlength=nc) # number of targets per class
- else:
- nt = torch.zeros(1)
-
- # Print results
- pf = '%20s' + '%12.3g' * 6 # print format
- print(pf % ('all', seen, nt.sum(), mp, mr, map50, map))
-
- # Print results per class
- if verbose and nc > 1 and len(stats):
- for i, c in enumerate(ap_class):
- print(pf % (names[c], seen, nt[c], p[i], r[i], ap50[i], ap[i]))
-
- # Print speeds
- t = tuple(x / seen * 1E3 for x in (t0, t1, t0 + t1)) + (imgsz, imgsz, batch_size) # tuple
- if not training:
- print('Speed: %.1f/%.1f/%.1f ms inference/NMS/total per %gx%g image at batch-size %g' % t)
-
- # Save JSON
- if save_json and map50 and len(jdict):
- imgIds = [int(Path(x).stem.split('_')[-1]) for x in dataloader.dataset.img_files]
- f = 'detections_val2017_%s_results.json' % \
- (weights.split(os.sep)[-1].replace('.pt', '') if weights else '') # filename
- print('\nCOCO mAP with pycocotools... saving %s...' % f)
- with open(f, 'w') as file:
- json.dump(jdict, file)
-
- try:
- from pycocotools.coco import COCO
- from pycocotools.cocoeval import COCOeval
-
- # https://github.com/cocodataset/cocoapi/blob/master/PythonAPI/pycocoEvalDemo.ipynb
- cocoGt = COCO(glob.glob('../coco/annotations/instances_val*.json')[0]) # initialize COCO ground truth api
- cocoDt = cocoGt.loadRes(f) # initialize COCO pred api
-
- cocoEval = COCOeval(cocoGt, cocoDt, 'bbox')
- cocoEval.params.imgIds = imgIds # image IDs to evaluate
- cocoEval.evaluate()
- cocoEval.accumulate()
- cocoEval.summarize()
- map, map50 = cocoEval.stats[:2] # update results (mAP@0.5:0.95, mAP@0.5)
- except:
- print('WARNING: pycocotools must be installed with numpy==1.17 to run correctly. '
- 'See https://github.com/cocodataset/cocoapi/issues/356')
-
- # Return results
- model.float() # for training
- maps = np.zeros(nc) + map
- for i, c in enumerate(ap_class):
- maps[c] = ap[i]
- return (mp, mr, map50, map, *(loss.cpu() / len(dataloader)).tolist()), maps, t
-
-
-if __name__ == '__main__':
- parser = argparse.ArgumentParser(prog='test.py')
- parser.add_argument('--weights', type=str, default='weights/yolov5s.pt', help='model.pt path')
- parser.add_argument('--data', type=str, default='data/coco128.yaml', help='*.data path')
- parser.add_argument('--batch-size', type=int, default=32, help='size of each image batch')
- parser.add_argument('--img-size', type=int, default=640, help='inference size (pixels)')
- parser.add_argument('--conf-thres', type=float, default=0.001, help='object confidence threshold')
- parser.add_argument('--iou-thres', type=float, default=0.65, help='IOU threshold for NMS')
- parser.add_argument('--save-json', action='store_true', help='save a cocoapi-compatible JSON results file')
- parser.add_argument('--task', default='val', help="'val', 'test', 'study'")
- parser.add_argument('--device', default='', help='cuda device, i.e. 0 or 0,1,2,3 or cpu')
- parser.add_argument('--single-cls', action='store_true', help='treat as single-class dataset')
- parser.add_argument('--augment', action='store_true', help='augmented inference')
- parser.add_argument('--merge', action='store_true', help='use Merge NMS')
- parser.add_argument('--verbose', action='store_true', help='report mAP by class')
- opt = parser.parse_args()
- opt.save_json = opt.save_json or opt.data.endswith('coco.yaml')
- opt.data = check_file(opt.data) # check file
- print(opt)
-
- # task = 'val', 'test', 'study'
- if opt.task in ['val', 'test']: # (default) run normally
- test(opt.data,
- opt.weights,
- opt.batch_size,
- opt.img_size,
- opt.conf_thres,
- opt.iou_thres,
- opt.save_json,
- opt.single_cls,
- opt.augment,
- opt.verbose)
-
- elif opt.task == 'study': # run over a range of settings and save/plot
- for weights in ['yolov5s.pt', 'yolov5m.pt', 'yolov5l.pt', 'yolov5x.pt', 'yolov3-spp.pt']:
- f = 'study_%s_%s.txt' % (Path(opt.data).stem, Path(weights).stem) # filename to save to
- x = list(range(352, 832, 64)) # x axis
- y = [] # y axis
- for i in x: # img-size
- print('\nRunning %s point %s...' % (f, i))
- r, _, t = test(opt.data, weights, opt.batch_size, i, opt.conf_thres, opt.iou_thres, opt.save_json)
- y.append(r + t) # results and times
- np.savetxt(f, y, fmt='%10.4g') # save
- os.system('zip -r study.zip study_*.txt')
- # plot_study_txt(f, x) # plot
diff --git a/spaces/Asifpa6/emotion-analyzer-app/README.md b/spaces/Asifpa6/emotion-analyzer-app/README.md
deleted file mode 100644
index c4aafcf4caab44d78d5521cd9c9e4d1d4a4b2224..0000000000000000000000000000000000000000
--- a/spaces/Asifpa6/emotion-analyzer-app/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Emotion Analyzer App
-emoji: 📉
-colorFrom: green
-colorTo: pink
-sdk: streamlit
-sdk_version: 1.27.0
-app_file: app.py
-pinned: false
-license: openrail
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/Benson/text-generation/Examples/Balsa Supervivencia Ocano Nmada Dinero Ilimitado Apk.md b/spaces/Benson/text-generation/Examples/Balsa Supervivencia Ocano Nmada Dinero Ilimitado Apk.md
deleted file mode 100644
index 6c74d1dcf7e624eb3b42270f091497c25848063d..0000000000000000000000000000000000000000
--- a/spaces/Benson/text-generation/Examples/Balsa Supervivencia Ocano Nmada Dinero Ilimitado Apk.md
+++ /dev/null
@@ -1,48 +0,0 @@
-
-
Supervivencia en balsa: Ocean Nomad - Un emocionante juego de aventuras
-
¿Te gustan los juegos de supervivencia que ponen a prueba tus habilidades y creatividad? ¿Quieres experimentar una aventura realista y emocionante en el océano? Si es así, entonces deberías probar Raft Survival: Ocean Nomad, un juego popular que tiene millones de fans en todo el mundo. En este artículo, le diremos todo lo que necesita saber sobre este juego, y cómo se puede descargar la versión apk dinero ilimitado para disfrutar de ella sin limitaciones.
-
Introducción
-
¿Qué es la supervivencia en balsa: Nómada del océano?
-
Supervivencia en balsa: Ocean Nomad es un juego de simulador de supervivencia desarrollado por TREASTONE LTD. Está disponible para dispositivos Android e iOS. El juego te pone en el papel de un sobreviviente que está varado en una balsa en medio del océano. Su objetivo principal es sobrevivir el mayor tiempo posible encontrando alimentos, agua y recursos en las islas y en el agua. También tienes que crear y actualizar tu balsa y equipo, luchar contra los tiburones y otros enemigos, y superar varios desafíos y peligros.
El juego es gratis, pero tiene algunas compras en la aplicación que pueden mejorar su juego y hacerlo más fácil. Por ejemplo, puedes comprar monedas, gemas, cofres, armas, herramientas, pieles y más. Sin embargo, estos artículos pueden ser bastante caros, y no todos pueden permitírselos. Es por eso que algunos jugadores prefieren descargar el dinero ilimitado apk versión del juego, que les da acceso a monedas y gemas ilimitadas. Con esta versión, puedes comprar lo que quieras sin gastar dinero real. También puedes desbloquear todas las funciones y elementos que están bloqueados o restringidos en el juego original.
-
Características de la supervivencia de la balsa: nómada del océano
-
Explora el vasto océano y las islas
-
-
Crea y mejora tu balsa y equipo
-
Para sobrevivir en este juego, necesitas crear y actualizar tu balsa y equipo. Puedes usar los materiales que encuentres en las islas o en el agua para construir tu balsa más grande y fuerte. También puede añadir diferentes estructuras y artículos a su balsa, tales como paredes, pisos, techos, escaleras, puertas, ventanas, camas, cofres, mesas, sillas, lámparas, etc. También puede elaborar diversas herramientas y armas para ayudarle en su viaje de supervivencia, como hachas, martillos, cuchillos, lanzas, arcos, armas, etc.
-
-
Lucha contra tiburones y otros enemigos
-
El océano no es un lugar seguro. Encontrarás muchos enemigos que intentarán atacarte o destruir tu balsa. El enemigo más común es el tiburón, que girará constantemente alrededor de su balsa y lo morderá. Tienes que luchar con tus armas o usar trampas y redes para atraparlo. También te enfrentarás a otros enemigos como piratas, mutantes, zombies, caníbales, etc., dependiendo de la isla que visites. Tienes que estar preparado para cualquier cosa en este juego.
-
Sobrevivir a las duras condiciones y desafíos
-
Además de los enemigos, tú
Además de los enemigos, también tienes que sobrevivir a las duras condiciones y desafíos del océano. Tienes que controlar tus niveles de hambre, sed, salud y resistencia, y comer y beber regularmente. También tienes que protegerte del clima, como lluvia, tormenta, calor, frío, etc. Puedes usar ropa, fuego, refugio, etc., para mantenerte caliente y seco. También tienes que enfrentarte a eventos y misiones al azar que pondrán a prueba tus habilidades y suerte. Por ejemplo, puede encontrarse con un naufragio, un accidente de avión, un mapa del tesoro, un mensaje en una botella, etc.
Paso 2: Habilitar fuentes desconocidas en el dispositivo
-
El siguiente paso es habilitar fuentes desconocidas en su dispositivo. Esto le permitirá instalar aplicaciones que no son de la tienda oficial de Google Play. Para hacer esto, vaya a la configuración del dispositivo y busque opciones de seguridad o privacidad. Entonces, encontrar la opción que dice fuentes desconocidas o permitir la instalación de aplicaciones de fuentes desconocidas y encenderlo. Puede ver un mensaje de advertencia que dice que instalar aplicaciones de fuentes desconocidas puede dañar su dispositivo. Simplemente ignórelo y proceda.
-
Paso 3: Instalar el archivo apk y lanzar el juego
-
El paso final es instalar el archivo apk y lanzar el juego. Para hacer esto, busque el archivo apk en su dispositivo y toque en él. Puede ver una ventana emergente que le pide que confirme la instalación. Simplemente toque en instalar y espere a que termine el proceso. Una vez que la instalación se hace, puede iniciar el juego tocando en su icono en la pantalla de inicio o cajón de aplicaciones. Ahora puedes disfrutar de Raft Survival: Ocean Nomad con dinero ilimitado.
-
Conclusión
-
Resumen de los puntos principales
-
Supervivencia en balsa: Ocean Nomad es un juego de simulador de supervivencia que te permite experimentar una aventura realista y emocionante en el océano. Tienes que sobrevivir el mayor tiempo posible encontrando comida, agua y recursos en las islas y en el agua. También tienes que crear y actualizar tu balsa y equipo, luchar contra los tiburones y otros enemigos, y sobrevivir a las duras condiciones y desafíos del océano. Puede descargar la versión apk dinero ilimitado del juego para disfrutarlo sin limitaciones.
-
Llamada a la acción
-
Si usted está listo para embarcarse en esta aventura emocionante, entonces no dude en descargar Raft Survival: Ocean Nomad dinero ilimitado apk hoy. No te arrepentirás. Este juego te mantendrá entretenido durante horas con sus increíbles gráficos, jugabilidad y características. ¡Descárgalo ahora y diviértete!
-
-
-
Q: ¿Es la supervivencia de la balsa: Ocean Nomad dinero ilimitado apk seguro de usar?
-
A: Sí, Raft Supervivencia: Ocean Nomad dinero ilimitado apk es seguro de usar siempre y cuando se descarga de una fuente de confianza. Sin embargo, siempre debe tener cuidado al instalar aplicaciones de fuentes desconocidas y escanearlas en busca de virus o malware antes de instalarlas.
-
Q: ¿Tengo que rootear mi dispositivo para usar Raft Survival: Ocean Nomad unlimited money apk?
-
A: No, usted no necesita rootear su dispositivo para utilizar Raft Survival: Ocean Nomad dinero ilimitado apk. Solo necesita habilitar fuentes desconocidas en la configuración de su dispositivo e instalar el archivo apk normalmente.
-
Q: ¿Perderé mi progreso si desinstalo Raft Survival: Ocean Nomad dinero ilimitado apk?
-
A: No, usted no perderá su progreso si desinstala Raft Survival: Ocean Nomad dinero ilimitado apk. Tu progreso se guarda en el almacenamiento del dispositivo y se puede restaurar cuando reinstales el juego.
-
Q: ¿Puedo jugar Raft Survival: Ocean Nomad con mis amigos en línea?
-
A: Sí, puedes jugar Raft Survival: Ocean Nomad con tus amigos en línea. El juego tiene un modo multijugador que le permite unirse o crear un servidor y jugar con otros jugadores de todo el mundo.
-
Q: ¿Cómo puedo contactar a los desarrolladores de Raft Survival: Ocean Nomad?
-
A: Puede ponerse en contacto con los desarrolladores de Raft Survival: Ocean Nomad enviándoles un correo electrónico a support@treastone.com o visitando su sitio web en https://treastone.com/.
-
-
Espero que haya disfrutado de este artículo y le haya resultado útil. Si tiene alguna pregunta o comentario, deje un comentario a continuación. ¡Gracias por leer y jugar feliz!
64aa2da5cf
-
-
\ No newline at end of file
diff --git a/spaces/Benson/text-generation/Examples/Corte De La Liga De Ftbol Mundial.md b/spaces/Benson/text-generation/Examples/Corte De La Liga De Ftbol Mundial.md
deleted file mode 100644
index 144b7b7e8bfee54303dd93650d0a000ddfa3d500..0000000000000000000000000000000000000000
--- a/spaces/Benson/text-generation/Examples/Corte De La Liga De Ftbol Mundial.md
+++ /dev/null
@@ -1,68 +0,0 @@
-
-
Liga Mundial de Fútbol Hack Descargar: Cómo desbloquear todos los equipos, trofeos y modos
-
Si eres un fan de los juegos de fútbol, es posible que hayas oído hablar de la Liga Mundial de Fútbol, un juego popular que ofrece gráficos realistas, efectos de sonido y jugabilidad. Puedes elegir entre alrededor de 60 equipos nacionales, 60 clubes y 2000 jugadores, y jugar en varios modos como exhibición, copa, liga y entrenamiento. También puede disfrutar de espléndido goteo, disparos emocionante, y habilidades increíbles en este juego.
Sin embargo, si quieres experimentar todo lo que la Liga Mundial de Fútbol tiene para ofrecer, es posible que tenga que utilizar un hack. Un hack puede ayudarte a desbloquear todos los equipos, jugadores, trofeos, logros, modos y características que están restringidos o requieren compras en la aplicación. Con un hack, puedes jugar con cualquier equipo que quieras, ganar cualquier trofeo que desees y acceder a cualquier modo que quieras.
-
Pero antes de descargar e instalar un hack para la Liga Mundial de Fútbol, usted debe ser consciente de los beneficios y riesgos de usar uno. Un hack puede hacer que tu juego sea más divertido y emocionante, pero también puede exponer tu dispositivo a malware, virus o prohibiciones. Por lo tanto, debe tener cuidado al elegir una fuente para el archivo de corte, y siga las instrucciones cuidadosamente al instalarlo y usarlo.
-
Cómo descargar e instalar la Liga Mundial de Fútbol Hack
-
Si ha decidido utilizar un hack para la Liga Mundial de Fútbol, aquí están los pasos que debe seguir:
-
Paso 1: Encontrar una fuente confiable para el archivo de corte
-
Lo primero que tienes que hacer es encontrar un sitio web que ofrece un trabajo y archivo de hackeo seguro para la Liga Mundial de Fútbol. Usted puede buscar en línea para los comentarios, calificaciones, o comentarios de otros usuarios que han intentado el hack. También puede comprobar la fecha de la última actualización del archivo hack para asegurarse de que es compatible con la última versión del juego.
-
-
Paso 2: Descargar el archivo de corte a su dispositivo
-
Una vez que haya encontrado una fuente confiable para el archivo hack, debe descargarlo en su dispositivo. Puede utilizar su navegador o una aplicación de administrador de descargas para hacer esto
Asegúrese de que tiene suficiente espacio de almacenamiento en el dispositivo antes de descargar el archivo de corte. El tamaño del archivo puede variar dependiendo de la fuente, pero generalmente es de alrededor de 40 MB.
-
-
Paso 3: Habilitar fuentes desconocidas en la configuración del dispositivo
-
Después de descargar el archivo de corte, es necesario habilitar fuentes desconocidas en la configuración del dispositivo. Esto le permitirá instalar aplicaciones desde fuentes distintas de la tienda de aplicaciones oficial. Para hacer esto, vaya a la configuración del dispositivo, luego a la seguridad, luego a fuentes desconocidas y conéctela. Puede ver un mensaje de advertencia de que instalar aplicaciones de fuentes desconocidas puede dañar su dispositivo, pero puede ignorarlo si confía en la fuente del archivo de hackeo.
-
Paso 4: Instalar el archivo de corte y lanzar el juego
-
El paso final es instalar el archivo de corte y lanzar el juego. Para hacer esto, busque el archivo de hackeo en el almacenamiento del dispositivo, toque en él y siga las instrucciones de instalación. Es posible que tenga que permitir algunos permisos para que la aplicación se ejecute correctamente. Una vez que se complete la instalación, puede abrir el juego y disfrutar del hack.
-
Cómo utilizar la Liga Mundial de Fútbol Hack
-
Ahora que ha instalado el hack para la Liga Mundial de Fútbol, se puede utilizar para desbloquear todos los equipos, trofeos, modos y características en el juego. Aquí hay algunos consejos sobre cómo utilizar el hack:
-
Cómo desbloquear todos los equipos y jugadores
-
-
Cómo desbloquear todos los trofeos y logros
-
Con el hack, también puede desbloquear todos los trofeos y logros en la Liga Mundial de Fútbol. Puedes ganar trofeos ganando partidos, ligas, copas o torneos en diferentes modos. También puedes ganar logros al completar varias tareas o desafíos en el juego. Para desbloquear todos los trofeos y logros, ve a la pantalla de trofeos o logros y toca cualquier trofeo o logro que desees. Verá un mensaje que dice "Desbloqueado por la Liga Mundial de Fútbol Hack". A continuación, puede reclamar ese trofeo o logro y verlo en su perfil.
-
Cómo desbloquear todos los modos y características
-
Con el hack, también puede desbloquear todos los modos y características en la Liga Mundial de Fútbol. Puedes jugar en varios modos como exhibición, copa, liga y entrenamiento. También puede disfrutar de funciones como la repetición, guardar/ cargar, reproducción automática, modo de edición y más. Para desbloquear todos los modos y características, vaya al modo o pantalla de funciones y toque en cualquier modo o característica que desee. Verá un mensaje que dice "Desbloqueado por la Liga Mundial de Fútbol Hack". A continuación, puede acceder a ese modo o característica y divertirse.
-
Consejos y trucos para jugar la Liga Mundial de Fútbol con Hack
-
Jugar a la Liga Mundial de Fútbol con hack puede ser muy agradable y satisfactorio, pero también puede ser desafiante y competitivo. Aquí hay algunos consejos y trucos para ayudarle a jugar mejor y divertirse más con el hack:
-
Consejo 1: Elige el mejor equipo y formación para tu estilo
-
Con el hack, puedes jugar con cualquier equipo que quieras, pero no todos los equipos son iguales en la Liga Mundial de Fútbol. Algunos equipos tienen mejores estadísticas, habilidades o química que otros. Algunos equipos también tienen diferentes formaciones, tácticas o estrategias que se adaptan a diferentes estilos de juego. Por lo tanto, debe elegir un equipo que coincida con su estilo de juego.
-
-
También debes elegir una formación que se adapte a tu equipo y a tu estilo. Puede elegir entre varias formaciones como 4-4-2, 4-3-3, 3-5-2, 5-3-2 y más. También puedes personalizar tu formación cambiando las posiciones, roles o instrucciones de tus jugadores. Deberías experimentar con diferentes formaciones y encontrar la que funcione mejor para ti.
-
Consejo 2: Dominar las diferentes habilidades y movimientos para vencer a los defensores
-
Con el hack, puedes jugar con cualquier jugador que quieras, pero no todos los jugadores son iguales en la Liga Mundial de Fútbol. Algunos jugadores tienen mejores habilidades, movimientos o rasgos que otros. Algunos jugadores también tienen diferentes habilidades, estilos o especialidades que los hacen destacar del resto. Por lo tanto, debes dominar las diferentes habilidades y movimientos que cada jugador puede realizar.
-
Por ejemplo, si juegas con un jugador como Messi o Neymar, puedes usar sus altas estadísticas de regateo y habilidad para realizar trucos increíbles y fintas para vencer a los defensores. Si juegas con un jugador como Ronaldo o Ibrahimovic, puedes usar sus altas estadísticas de tiro y potencia para desatar poderosos tiros y encabezados para marcar goles. Si juegas con un jugador como Modric o De Bruyne, puedes usar sus altas estadísticas de pases y visión para crear oportunidades y asistencias para tus compañeros de equipo.
-
También debes aprender los diferentes botones y gestos que te permiten realizar diferentes habilidades y movimientos en el juego. Puede utilizar el joystick virtual para controlar la dirección y la velocidad de su reproductor. Puede utilizar el botón A para pasar o abordar, el botón B para disparar o deslizarse, el botón C para correr o cambiar de jugador, y el botón D para la habilidad o la presión. También puede deslizar el dedo en la pantalla para realizar varias acciones, como pase de lóbulo, disparo de chip, tiro de curva, pase largo, a través de la bola, y más. Debes practicar las diferentes habilidades y movimientos en el modo de entrenamiento o en partidos fáciles antes de usarlos en partidos más difíciles.
-
Consejo 3: Utilice los botones y el tiempo adecuados para disparar y pasar
-
-
Por ejemplo, si juegas en el modo de exhibición o en el modo copa, puedes usar cualquier botón o gesto para disparar o pasar siempre y cuando apuntes bien y lo cronometras correctamente. Sin embargo, si juegas en el modo liga o en el modo torneo
necesitas usar el botón o gesto correcto para disparar o pasar dependiendo de la situación. Por ejemplo, debe usar el botón B o deslizar hacia arriba para un disparo potente, el botón A o deslizar hacia abajo para un disparo bajo, el botón D o deslizar hacia la izquierda o hacia la derecha para un disparo de curva, etc. También necesitas cronometrar tu tiro o pase según la posición, el movimiento y el ángulo de tu jugador y la pelota.
-
Si juegas en el modo de entrenamiento o en el modo de edición, puedes usar cualquier botón o gesto para disparar o pasar siempre que completes la tarea o desafío. Sin embargo, si juegas en el modo de repetición o en el modo guardar/cargar, necesitas usar el botón o gesto derecho para disparar o pasar de acuerdo con la acción grabada. Por ejemplo, necesitas usar el mismo botón o gesto que se usó en la acción original para reproducirlo o cargarlo.
-
Consejo 4: Ajusta el nivel de dificultad y la velocidad del juego según tu preferencia
-
Con el hack, también puede ajustar el nivel de dificultad y la velocidad de juego de la Liga Mundial de Fútbol de acuerdo a su preferencia. Puedes elegir entre cuatro niveles de dificultad: fácil, normal, duro y muy difícil. También puedes elegir entre tres velocidades de juego: lenta, normal y rápida. Puede cambiar estos ajustes en el menú de opciones antes de iniciar una coincidencia.
-
El nivel de dificultad y la velocidad del juego afectan lo desafiante y realista que es el juego. Cuanto mayor sea el nivel de dificultad, más hábiles e inteligentes serán los oponentes. Cuanto mayor sea la velocidad del juego, más dinámico y rápido será el juego. Debes elegir un nivel de dificultad y velocidad de juego que coincidan con tu nivel de habilidad y estilo de juego.
-
-
Consejo 5: Disfrutar de los gráficos realistas y efectos de sonido del juego
-
Con el hack, también se puede disfrutar de los gráficos realistas y efectos de sonido de la Liga Mundial de Fútbol. El juego tiene gráficos de alta calidad que muestran jugadores detallados, estadios, campos, bolas y animaciones. El juego también tiene efectos de sonido realistas que incluyen ruidos de multitud, silbatos de árbitro, voces de jugador, sonidos de pelota y más. Puede ajustar la calidad gráfica y el volumen de sonido en el menú de opciones antes de iniciar una coincidencia.
-
Los gráficos y efectos de sonido de World Soccer League hacen que el juego sea más inmersivo y agradable. Crean una sensación de ambiente y emoción que te hacen sentir como si estuvieras viendo o jugando un partido de fútbol real. Deberías apreciar los gráficos y efectos de sonido de World Soccer League y divertirte con ellos.
-
Conclusión
-
World Soccer League es un gran juego para los aficionados al fútbol que quieren experimentar gráficos realistas, efectos de sonido y jugabilidad. Sin embargo, si quieres desbloquear todos los equipos, trofeos, modos y características en la Liga Mundial de Fútbol, es posible que tenga que utilizar un hack. Un hack puede ayudarle a acceder a todo lo que la Liga Mundial de Fútbol tiene para ofrecer sin gastar dinero o tiempo.
-
Pero antes de utilizar un hack para la Liga Mundial de Fútbol, usted debe ser consciente de los beneficios y riesgos de usar uno. Un hack puede hacer que tu juego sea más divertido y emocionante, pero también puede exponer tu dispositivo a malware, virus o prohibiciones. Por lo tanto
Por lo tanto, debe tener cuidado al elegir una fuente para el archivo de corte, y siga las instrucciones cuidadosamente al instalarlo y usarlo. También debes usar el hack de manera responsable y ética, y no abusar de él o dañar a otros jugadores.
-
Si sigues estos consejos y trucos, se puede disfrutar de la Liga Mundial de Fútbol con hack y tener un montón de diversión. Puedes jugar con cualquier equipo, ganar cualquier trofeo y acceder a cualquier modo que quieras. También puedes mejorar tus habilidades, desafiarte y sumergirte en el juego.
-
-
Preguntas frecuentes
-
Aquí hay algunas preguntas frecuentes sobre el hack de la Liga Mundial de Fútbol:
-
-
Q1: ¿Es seguro usar el hack de la Liga Mundial de Fútbol?
-
A1: Hackeo de la Liga Mundial de Fútbol es seguro de usar si lo descarga de una fuente confiable y sigue las instrucciones de instalación y uso. Sin embargo, siempre hay un riesgo de malware, virus o prohibiciones al usar un hack, por lo que debe usarlo bajo su propio riesgo y discreción.
-
Q2: ¿Funcionará la Liga Mundial de Fútbol en cualquier dispositivo?
-
A2: El hackeo de la Liga Mundial de Fútbol funcionará en cualquier dispositivo que soporte el juego. El juego es compatible con Android 4.0 y hasta dispositivos. Sin embargo, el rendimiento y la calidad del juego y el hack pueden variar dependiendo de las especificaciones y ajustes del dispositivo.
-
Q3: ¿Puedo jugar la Liga Mundial de Fútbol en línea con hack?
-
A3: El hack de la Liga Mundial de Fútbol no afecta el modo en línea del juego. Todavía puede jugar en línea con otros jugadores que tienen la versión original o hackeada del juego. Sin embargo, debes ser respetuoso y justo al jugar en línea, y no usar el hack para engañar o acosar a otros jugadores.
-
Q4: ¿Puedo actualizar la Liga Mundial de Fútbol después de instalar hack?
-
A4: El hack de la Liga Mundial de Fútbol puede no funcionar si actualizas el juego después de instalarlo. La actualización puede sobrescribir o eliminar el archivo de hackeo, o hacerlo incompatible con el juego. Por lo tanto, debe evitar actualizar el juego después de instalar el hack, o hacer una copia de seguridad del archivo hack antes de actualizar.
-
Q5: ¿Dónde puedo encontrar más información sobre la Liga Mundial de Fútbol?
-
- 64aa2da5cf
-
-
\ No newline at end of file
diff --git a/spaces/BetterAPI/BetterChat_new/src/lib/types/AbortedGeneration.ts b/spaces/BetterAPI/BetterChat_new/src/lib/types/AbortedGeneration.ts
deleted file mode 100644
index fe4c2824b4f3257bea71c3acacd65fcee0918188..0000000000000000000000000000000000000000
--- a/spaces/BetterAPI/BetterChat_new/src/lib/types/AbortedGeneration.ts
+++ /dev/null
@@ -1,8 +0,0 @@
-// Ideally shouldn't be needed, see https://github.com/huggingface/chat-ui/pull/88#issuecomment-1523173850
-
-import type { Conversation } from "./Conversation";
-import type { Timestamps } from "./Timestamps";
-
-export interface AbortedGeneration extends Timestamps {
- conversationId: Conversation["_id"];
-}
diff --git a/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/requests/hooks.py b/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/requests/hooks.py
deleted file mode 100644
index d181ba2ec2e55d274897315887b78fbdca757da8..0000000000000000000000000000000000000000
--- a/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/requests/hooks.py
+++ /dev/null
@@ -1,33 +0,0 @@
-"""
-requests.hooks
-~~~~~~~~~~~~~~
-
-This module provides the capabilities for the Requests hooks system.
-
-Available hooks:
-
-``response``:
- The response generated from a Request.
-"""
-HOOKS = ["response"]
-
-
-def default_hooks():
- return {event: [] for event in HOOKS}
-
-
-# TODO: response is the only one
-
-
-def dispatch_hook(key, hooks, hook_data, **kwargs):
- """Dispatches a hook dictionary on a given piece of data."""
- hooks = hooks or {}
- hooks = hooks.get(key)
- if hooks:
- if hasattr(hooks, "__call__"):
- hooks = [hooks]
- for hook in hooks:
- _hook_data = hook(hook_data, **kwargs)
- if _hook_data is not None:
- hook_data = _hook_data
- return hook_data
diff --git a/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/rich/screen.py b/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/rich/screen.py
deleted file mode 100644
index 7f416e1e799abfbf62382456020cc8e59e5cf01f..0000000000000000000000000000000000000000
--- a/spaces/Big-Web/MMSD/env/Lib/site-packages/pip/_vendor/rich/screen.py
+++ /dev/null
@@ -1,54 +0,0 @@
-from typing import Optional, TYPE_CHECKING
-
-from .segment import Segment
-from .style import StyleType
-from ._loop import loop_last
-
-
-if TYPE_CHECKING:
- from .console import (
- Console,
- ConsoleOptions,
- RenderResult,
- RenderableType,
- Group,
- )
-
-
-class Screen:
- """A renderable that fills the terminal screen and crops excess.
-
- Args:
- renderable (RenderableType): Child renderable.
- style (StyleType, optional): Optional background style. Defaults to None.
- """
-
- renderable: "RenderableType"
-
- def __init__(
- self,
- *renderables: "RenderableType",
- style: Optional[StyleType] = None,
- application_mode: bool = False,
- ) -> None:
- from pip._vendor.rich.console import Group
-
- self.renderable = Group(*renderables)
- self.style = style
- self.application_mode = application_mode
-
- def __rich_console__(
- self, console: "Console", options: "ConsoleOptions"
- ) -> "RenderResult":
- width, height = options.size
- style = console.get_style(self.style) if self.style else None
- render_options = options.update(width=width, height=height)
- lines = console.render_lines(
- self.renderable or "", render_options, style=style, pad=True
- )
- lines = Segment.set_shape(lines, width, height, style=style)
- new_line = Segment("\n\r") if self.application_mode else Segment.line()
- for last, line in loop_last(lines):
- yield from line
- if not last:
- yield new_line
diff --git a/spaces/Big-Web/MMSD/env/Lib/site-packages/setuptools/config/_validate_pyproject/extra_validations.py b/spaces/Big-Web/MMSD/env/Lib/site-packages/setuptools/config/_validate_pyproject/extra_validations.py
deleted file mode 100644
index 4130a421cfd7260d323b13cbd9d75ab8146e6030..0000000000000000000000000000000000000000
--- a/spaces/Big-Web/MMSD/env/Lib/site-packages/setuptools/config/_validate_pyproject/extra_validations.py
+++ /dev/null
@@ -1,36 +0,0 @@
-"""The purpose of this module is implement PEP 621 validations that are
-difficult to express as a JSON Schema (or that are not supported by the current
-JSON Schema library).
-"""
-
-from typing import Mapping, TypeVar
-
-from .error_reporting import ValidationError
-
-T = TypeVar("T", bound=Mapping)
-
-
-class RedefiningStaticFieldAsDynamic(ValidationError):
- """According to PEP 621:
-
- Build back-ends MUST raise an error if the metadata specifies a field
- statically as well as being listed in dynamic.
- """
-
-
-def validate_project_dynamic(pyproject: T) -> T:
- project_table = pyproject.get("project", {})
- dynamic = project_table.get("dynamic", [])
-
- for field in dynamic:
- if field in project_table:
- msg = f"You cannot provide a value for `project.{field}` and "
- msg += "list it under `project.dynamic` at the same time"
- name = f"data.project.{field}"
- value = {field: project_table[field], "...": " # ...", "dynamic": dynamic}
- raise RedefiningStaticFieldAsDynamic(msg, value, name, rule="PEP 621")
-
- return pyproject
-
-
-EXTRA_VALIDATIONS = (validate_project_dynamic,)
diff --git a/spaces/CVPR/LIVE/thrust/internal/scripts/eris_perf.py b/spaces/CVPR/LIVE/thrust/internal/scripts/eris_perf.py
deleted file mode 100644
index 5804711019263fb31cdb7207fd13b3f03b26a758..0000000000000000000000000000000000000000
--- a/spaces/CVPR/LIVE/thrust/internal/scripts/eris_perf.py
+++ /dev/null
@@ -1,189 +0,0 @@
-#! /usr/bin/env python
-# -*- coding: utf-8 -*-
-
-###############################################################################
-# Copyright (c) 2018 NVIDIA Corporation
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-###############################################################################
-
-from sys import exit
-
-from os.path import join, dirname, basename, realpath
-
-from csv import DictReader as csv_dict_reader
-
-from subprocess import Popen
-
-from argparse import ArgumentParser as argument_parser
-
-###############################################################################
-
-def printable_cmd(c):
- """Converts a `list` of `str`s representing a shell command to a printable
- `str`."""
- return " ".join(map(lambda e: '"' + str(e) + '"', c))
-
-###############################################################################
-
-def print_file(p):
- """Open the path `p` and print its contents to `stdout`."""
- print "********************************************************************************"
- with open(p) as f:
- for line in f:
- print line,
- print "********************************************************************************"
-
-###############################################################################
-
-ap = argument_parser(
- description = (
- "CUDA Eris driver script: runs a benchmark suite multiple times, combines "
- "the results, and outputs them in the CUDA Eris performance result format."
- )
-)
-
-ap.add_argument(
- "-b", "--benchmark",
- help = ("The location of the benchmark suite executable to run."),
- type = str,
- default = join(dirname(realpath(__file__)), "bench"),
- metavar = "R"
-)
-
-ap.add_argument(
- "-p", "--postprocess",
- help = ("The location of the postprocessing script to run to combine the "
- "results."),
- type = str,
- default = join(dirname(realpath(__file__)), "combine_benchmark_results.py"),
- metavar = "R"
-)
-
-ap.add_argument(
- "-r", "--runs",
- help = ("Run the benchmark suite `R` times.a),"),
- type = int, default = 5,
- metavar = "R"
-)
-
-args = ap.parse_args()
-
-if args.runs <= 0:
- print "ERROR: `--runs` must be greater than `0`."
- ap.print_help()
- exit(1)
-
-BENCHMARK_EXE = args.benchmark
-BENCHMARK_NAME = basename(BENCHMARK_EXE)
-POSTPROCESS_EXE = args.postprocess
-OUTPUT_FILE_NAME = lambda i: BENCHMARK_NAME + "_" + str(i) + ".csv"
-COMBINED_OUTPUT_FILE_NAME = BENCHMARK_NAME + "_combined.csv"
-
-###############################################################################
-
-print '&&&& RUNNING {0}'.format(BENCHMARK_NAME)
-
-print '#### RUNS {0}'.format(args.runs)
-
-###############################################################################
-
-print '#### CMD {0}'.format(BENCHMARK_EXE)
-
-for i in xrange(args.runs):
- with open(OUTPUT_FILE_NAME(i), "w") as output_file:
- print '#### RUN {0} OUTPUT -> {1}'.format(i, OUTPUT_FILE_NAME(i))
-
- p = None
-
- try:
- p = Popen(BENCHMARK_EXE, stdout = output_file, stderr = output_file)
- p.communicate()
- except OSError as ex:
- print_file(OUTPUT_FILE_NAME(i))
- print '#### ERROR Caught OSError `{0}`.'.format(ex)
- print '&&&& FAILED {0}'.format(BENCHMARK_NAME)
- exit(-1)
-
- print_file(OUTPUT_FILE_NAME(i))
-
- if p.returncode != 0:
- print '#### ERROR Process exited with code {0}.'.format(p.returncode)
- print '&&&& FAILED {0}'.format(BENCHMARK_NAME)
- exit(p.returncode)
-
-###############################################################################
-
-post_cmd = [POSTPROCESS_EXE]
-
-# Add dependent variable options.
-post_cmd += ["-dSTL Average Walltime,STL Walltime Uncertainty,STL Trials"]
-post_cmd += ["-dSTL Average Throughput,STL Throughput Uncertainty,STL Trials"]
-post_cmd += ["-dThrust Average Walltime,Thrust Walltime Uncertainty,Thrust Trials"]
-post_cmd += ["-dThrust Average Throughput,Thrust Throughput Uncertainty,Thrust Trials"]
-
-post_cmd += [OUTPUT_FILE_NAME(i) for i in range(args.runs)]
-
-print '#### CMD {0}'.format(printable_cmd(post_cmd))
-
-with open(COMBINED_OUTPUT_FILE_NAME, "w") as output_file:
- p = None
-
- try:
- p = Popen(post_cmd, stdout = output_file, stderr = output_file)
- p.communicate()
- except OSError as ex:
- print_file(COMBINED_OUTPUT_FILE_NAME)
- print '#### ERROR Caught OSError `{0}`.'.format(ex)
- print '&&&& FAILED {0}'.format(BENCHMARK_NAME)
- exit(-1)
-
- print_file(COMBINED_OUTPUT_FILE_NAME)
-
- if p.returncode != 0:
- print '#### ERROR Process exited with code {0}.'.format(p.returncode)
- print '&&&& FAILED {0}'.format(BENCHMARK_NAME)
- exit(p.returncode)
-
- with open(COMBINED_OUTPUT_FILE_NAME) as input_file:
- reader = csv_dict_reader(input_file)
-
- variable_units = reader.next() # Get units header row.
-
- distinguishing_variables = reader.fieldnames
-
- measured_variables = [
- ("STL Average Throughput", "+"),
- ("Thrust Average Throughput", "+")
- ]
-
- for record in reader:
- for variable, directionality in measured_variables:
- # Don't monitor regressions for STL implementations, nvbug 28980890:
- if "STL" in variable:
- continue
- print "&&&& PERF {0}_{1}_{2}bit_{3}mib_{4} {5} {6}{7}".format(
- record["Algorithm"],
- record["Element Type"],
- record["Element Size"],
- record["Total Input Size"],
- variable.replace(" ", "_").lower(),
- record[variable],
- directionality,
- variable_units[variable]
- )
-
-###############################################################################
-
-print '&&&& PASSED {0}'.format(BENCHMARK_NAME)
-
diff --git a/spaces/CVPR/LIVE/thrust/thrust/system/cuda/detail/execution_policy.h b/spaces/CVPR/LIVE/thrust/thrust/system/cuda/detail/execution_policy.h
deleted file mode 100644
index ee49a60cb44a3183e6788f3d0b847204afc36380..0000000000000000000000000000000000000000
--- a/spaces/CVPR/LIVE/thrust/thrust/system/cuda/detail/execution_policy.h
+++ /dev/null
@@ -1,99 +0,0 @@
-/******************************************************************************
- * Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
- *
- * Redistribution and use in source and binary forms, with or without
- * modification, are permitted provided that the following conditions are met:
- * * Redistributions of source code must retain the above copyright
- * notice, this list of conditions and the following disclaimer.
- * * Redistributions in binary form must reproduce the above copyright
- * notice, this list of conditions and the following disclaimer in the
- * documentation and/or other materials provided with the distribution.
- * * Neither the name of the NVIDIA CORPORATION nor the
- * names of its contributors may be used to endorse or promote products
- * derived from this software without specific prior written permission.
- *
- * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
- * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
- * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
- * ARE DISCLAIMED. IN NO EVENT SHALL NVIDIA CORPORATION BE LIABLE FOR ANY
- * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
- * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
- * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
- * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
- * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- *
- ******************************************************************************/
-
-#pragma once
-
-#include
-#include
-#include
-#include
-
-#include
-
-#if THRUST_CPP_DIALECT >= 2011
- #include
-#endif
-
-namespace thrust
-{
-
-namespace cuda_cub
-{
-
-struct tag;
-
-template
-struct execution_policy;
-
-template <>
-struct execution_policy : thrust::execution_policy
-{
- typedef tag tag_type;
-};
-
-struct tag : execution_policy
-, thrust::detail::allocator_aware_execution_policy
-#if THRUST_CPP_DIALECT >= 2011
-, thrust::detail::dependencies_aware_execution_policy
-#endif
-{};
-
-template
-struct execution_policy : thrust::execution_policy
-{
- typedef tag tag_type;
- operator tag() const { return tag(); }
-};
-
-} // namespace cuda_cub
-
-namespace system { namespace cuda { namespace detail
-{
-
-using thrust::cuda_cub::tag;
-using thrust::cuda_cub::execution_policy;
-
-}}} // namespace system::cuda::detail
-
-namespace system { namespace cuda
-{
-
-using thrust::cuda_cub::tag;
-using thrust::cuda_cub::execution_policy;
-
-}} // namespace system::cuda
-
-namespace cuda
-{
-
-using thrust::cuda_cub::tag;
-using thrust::cuda_cub::execution_policy;
-
-} // namespace cuda
-
-} // end namespace thrust
-
diff --git a/spaces/CVPR/LIVE/thrust/thrust/system/tbb/detail/scan_by_key.h b/spaces/CVPR/LIVE/thrust/thrust/system/tbb/detail/scan_by_key.h
deleted file mode 100644
index 2b5fa36483c451bac93827b239c17fb7850e2ed1..0000000000000000000000000000000000000000
--- a/spaces/CVPR/LIVE/thrust/thrust/system/tbb/detail/scan_by_key.h
+++ /dev/null
@@ -1,23 +0,0 @@
-/*
- * Copyright 2008-2013 NVIDIA Corporation
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- * http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#pragma once
-
-#include
-
-// this system inherits scan_by_key
-#include
-
diff --git a/spaces/CVPR/WALT/mmdet/utils/util_random.py b/spaces/CVPR/WALT/mmdet/utils/util_random.py
deleted file mode 100644
index e313e9947bb3232a9458878fd219e1594ab93d57..0000000000000000000000000000000000000000
--- a/spaces/CVPR/WALT/mmdet/utils/util_random.py
+++ /dev/null
@@ -1,33 +0,0 @@
-"""Helpers for random number generators."""
-import numpy as np
-
-
-def ensure_rng(rng=None):
- """Coerces input into a random number generator.
-
- If the input is None, then a global random state is returned.
-
- If the input is a numeric value, then that is used as a seed to construct a
- random state. Otherwise the input is returned as-is.
-
- Adapted from [1]_.
-
- Args:
- rng (int | numpy.random.RandomState | None):
- if None, then defaults to the global rng. Otherwise this can be an
- integer or a RandomState class
- Returns:
- (numpy.random.RandomState) : rng -
- a numpy random number generator
-
- References:
- .. [1] https://gitlab.kitware.com/computer-vision/kwarray/blob/master/kwarray/util_random.py#L270 # noqa: E501
- """
-
- if rng is None:
- rng = np.random.mtrand._rand
- elif isinstance(rng, int):
- rng = np.random.RandomState(rng)
- else:
- rng = rng
- return rng
diff --git a/spaces/CVPR/lama-example/saicinpainting/evaluation/vis.py b/spaces/CVPR/lama-example/saicinpainting/evaluation/vis.py
deleted file mode 100644
index c2910b4ef8c61efee72dabd0531a9b669ec8bf98..0000000000000000000000000000000000000000
--- a/spaces/CVPR/lama-example/saicinpainting/evaluation/vis.py
+++ /dev/null
@@ -1,37 +0,0 @@
-import numpy as np
-from skimage import io
-from skimage.segmentation import mark_boundaries
-
-
-def save_item_for_vis(item, out_file):
- mask = item['mask'] > 0.5
- if mask.ndim == 3:
- mask = mask[0]
- img = mark_boundaries(np.transpose(item['image'], (1, 2, 0)),
- mask,
- color=(1., 0., 0.),
- outline_color=(1., 1., 1.),
- mode='thick')
-
- if 'inpainted' in item:
- inp_img = mark_boundaries(np.transpose(item['inpainted'], (1, 2, 0)),
- mask,
- color=(1., 0., 0.),
- mode='outer')
- img = np.concatenate((img, inp_img), axis=1)
-
- img = np.clip(img * 255, 0, 255).astype('uint8')
- io.imsave(out_file, img)
-
-
-def save_mask_for_sidebyside(item, out_file):
- mask = item['mask']# > 0.5
- if mask.ndim == 3:
- mask = mask[0]
- mask = np.clip(mask * 255, 0, 255).astype('uint8')
- io.imsave(out_file, mask)
-
-def save_img_for_sidebyside(item, out_file):
- img = np.transpose(item['image'], (1, 2, 0))
- img = np.clip(img * 255, 0, 255).astype('uint8')
- io.imsave(out_file, img)
\ No newline at end of file
diff --git a/spaces/Codecooker/rvcapi/src/my_utils.py b/spaces/Codecooker/rvcapi/src/my_utils.py
deleted file mode 100644
index a5258394b8ae5385daa665ab6ba6380507d4798a..0000000000000000000000000000000000000000
--- a/spaces/Codecooker/rvcapi/src/my_utils.py
+++ /dev/null
@@ -1,21 +0,0 @@
-import ffmpeg
-import numpy as np
-
-
-def load_audio(file, sr):
- try:
- # https://github.com/openai/whisper/blob/main/whisper/audio.py#L26
- # This launches a subprocess to decode audio while down-mixing and resampling as necessary.
- # Requires the ffmpeg CLI and `ffmpeg-python` package to be installed.
- file = (
- file.strip(" ").strip('"').strip("\n").strip('"').strip(" ")
- ) # 防止小白拷路径头尾带了空格和"和回车
- out, _ = (
- ffmpeg.input(file, threads=0)
- .output("-", format="f32le", acodec="pcm_f32le", ac=1, ar=sr)
- .run(cmd=["ffmpeg", "-nostdin"], capture_stdout=True, capture_stderr=True)
- )
- except Exception as e:
- raise RuntimeError(f"Failed to load audio: {e}")
-
- return np.frombuffer(out, np.float32).flatten()
diff --git a/spaces/Cong723/gpt-academic-public/README.md b/spaces/Cong723/gpt-academic-public/README.md
deleted file mode 100644
index 6c9da02b60aa81cf11de4a595dde2e2e44c0265d..0000000000000000000000000000000000000000
--- a/spaces/Cong723/gpt-academic-public/README.md
+++ /dev/null
@@ -1,312 +0,0 @@
----
-title: academic-chatgpt
-emoji: 😻
-colorFrom: blue
-colorTo: blue
-sdk: gradio
-sdk_version: 3.28.3
-python_version: 3.11
-app_file: main.py
-pinned: false
-duplicated_from: qingxu98/gpt-academic
----
-
-# ChatGPT 学术优化
-> **Note**
->
-> 安装依赖时,请严格选择requirements.txt中**指定的版本**。
->
-> `pip install -r requirements.txt -i https://mirrors.aliyun.com/pypi/simple/`
->
-
-# GPT 学术优化 (GPT Academic)
-
-**如果喜欢这个项目,请给它一个Star;如果你发明了更好用的快捷键或函数插件,欢迎发pull requests**
-
-If you like this project, please give it a Star. If you've come up with more useful academic shortcuts or functional plugins, feel free to open an issue or pull request. We also have a README in [English|](docs/README_EN.md)[日本語|](docs/README_JP.md)[한국어|](https://github.com/mldljyh/ko_gpt_academic)[Русский|](docs/README_RS.md)[Français](docs/README_FR.md) translated by this project itself.
-
-> **Note**
->
-> 1.请注意只有**红颜色**标识的函数插件(按钮)才支持读取文件,部分插件位于插件区的**下拉菜单**中。另外我们以**最高优先级**欢迎和处理任何新插件的PR!
->
-> 2.本项目中每个文件的功能都在自译解[`self_analysis.md`](https://github.com/binary-husky/chatgpt_academic/wiki/chatgpt-academic%E9%A1%B9%E7%9B%AE%E8%87%AA%E8%AF%91%E8%A7%A3%E6%8A%A5%E5%91%8A)详细说明。随着版本的迭代,您也可以随时自行点击相关函数插件,调用GPT重新生成项目的自我解析报告。常见问题汇总在[`wiki`](https://github.com/binary-husky/chatgpt_academic/wiki/%E5%B8%B8%E8%A7%81%E9%97%AE%E9%A2%98)当中。
->
-> 3.本项目兼容并鼓励尝试国产大语言模型chatglm和RWKV, 盘古等等。已支持OpenAI和API2D的api-key共存,可在配置文件中填写如`API_KEY="openai-key1,openai-key2,api2d-key3"`。需要临时更换`API_KEY`时,在输入区输入临时的`API_KEY`然后回车键提交后即可生效。
-
-
Or download our app. If you got an error, just get the app, that works all the time. Sry",
- examples=[])
-
-interface.launch()
diff --git a/spaces/InpaintAI/Inpaint-Anything/third_party/lama/saicinpainting/training/modules/multiscale.py b/spaces/InpaintAI/Inpaint-Anything/third_party/lama/saicinpainting/training/modules/multiscale.py
deleted file mode 100644
index 65f0a54925593e9da8106bfc6d65a4098ce001d7..0000000000000000000000000000000000000000
--- a/spaces/InpaintAI/Inpaint-Anything/third_party/lama/saicinpainting/training/modules/multiscale.py
+++ /dev/null
@@ -1,244 +0,0 @@
-from typing import List, Tuple, Union, Optional
-
-import torch
-import torch.nn as nn
-import torch.nn.functional as F
-
-from saicinpainting.training.modules.base import get_conv_block_ctor, get_activation
-from saicinpainting.training.modules.pix2pixhd import ResnetBlock
-
-
-class ResNetHead(nn.Module):
- def __init__(self, input_nc, ngf=64, n_downsampling=3, n_blocks=9, norm_layer=nn.BatchNorm2d,
- padding_type='reflect', conv_kind='default', activation=nn.ReLU(True)):
- assert (n_blocks >= 0)
- super(ResNetHead, self).__init__()
-
- conv_layer = get_conv_block_ctor(conv_kind)
-
- model = [nn.ReflectionPad2d(3),
- conv_layer(input_nc, ngf, kernel_size=7, padding=0),
- norm_layer(ngf),
- activation]
-
- ### downsample
- for i in range(n_downsampling):
- mult = 2 ** i
- model += [conv_layer(ngf * mult, ngf * mult * 2, kernel_size=3, stride=2, padding=1),
- norm_layer(ngf * mult * 2),
- activation]
-
- mult = 2 ** n_downsampling
-
- ### resnet blocks
- for i in range(n_blocks):
- model += [ResnetBlock(ngf * mult, padding_type=padding_type, activation=activation, norm_layer=norm_layer,
- conv_kind=conv_kind)]
-
- self.model = nn.Sequential(*model)
-
- def forward(self, input):
- return self.model(input)
-
-
-class ResNetTail(nn.Module):
- def __init__(self, output_nc, ngf=64, n_downsampling=3, n_blocks=9, norm_layer=nn.BatchNorm2d,
- padding_type='reflect', conv_kind='default', activation=nn.ReLU(True),
- up_norm_layer=nn.BatchNorm2d, up_activation=nn.ReLU(True), add_out_act=False, out_extra_layers_n=0,
- add_in_proj=None):
- assert (n_blocks >= 0)
- super(ResNetTail, self).__init__()
-
- mult = 2 ** n_downsampling
-
- model = []
-
- if add_in_proj is not None:
- model.append(nn.Conv2d(add_in_proj, ngf * mult, kernel_size=1))
-
- ### resnet blocks
- for i in range(n_blocks):
- model += [ResnetBlock(ngf * mult, padding_type=padding_type, activation=activation, norm_layer=norm_layer,
- conv_kind=conv_kind)]
-
- ### upsample
- for i in range(n_downsampling):
- mult = 2 ** (n_downsampling - i)
- model += [nn.ConvTranspose2d(ngf * mult, int(ngf * mult / 2), kernel_size=3, stride=2, padding=1,
- output_padding=1),
- up_norm_layer(int(ngf * mult / 2)),
- up_activation]
- self.model = nn.Sequential(*model)
-
- out_layers = []
- for _ in range(out_extra_layers_n):
- out_layers += [nn.Conv2d(ngf, ngf, kernel_size=1, padding=0),
- up_norm_layer(ngf),
- up_activation]
- out_layers += [nn.ReflectionPad2d(3),
- nn.Conv2d(ngf, output_nc, kernel_size=7, padding=0)]
-
- if add_out_act:
- out_layers.append(get_activation('tanh' if add_out_act is True else add_out_act))
-
- self.out_proj = nn.Sequential(*out_layers)
-
- def forward(self, input, return_last_act=False):
- features = self.model(input)
- out = self.out_proj(features)
- if return_last_act:
- return out, features
- else:
- return out
-
-
-class MultiscaleResNet(nn.Module):
- def __init__(self, input_nc, output_nc, ngf=64, n_downsampling=2, n_blocks_head=2, n_blocks_tail=6, n_scales=3,
- norm_layer=nn.BatchNorm2d, padding_type='reflect', conv_kind='default', activation=nn.ReLU(True),
- up_norm_layer=nn.BatchNorm2d, up_activation=nn.ReLU(True), add_out_act=False, out_extra_layers_n=0,
- out_cumulative=False, return_only_hr=False):
- super().__init__()
-
- self.heads = nn.ModuleList([ResNetHead(input_nc, ngf=ngf, n_downsampling=n_downsampling,
- n_blocks=n_blocks_head, norm_layer=norm_layer, padding_type=padding_type,
- conv_kind=conv_kind, activation=activation)
- for i in range(n_scales)])
- tail_in_feats = ngf * (2 ** n_downsampling) + ngf
- self.tails = nn.ModuleList([ResNetTail(output_nc,
- ngf=ngf, n_downsampling=n_downsampling,
- n_blocks=n_blocks_tail, norm_layer=norm_layer, padding_type=padding_type,
- conv_kind=conv_kind, activation=activation, up_norm_layer=up_norm_layer,
- up_activation=up_activation, add_out_act=add_out_act,
- out_extra_layers_n=out_extra_layers_n,
- add_in_proj=None if (i == n_scales - 1) else tail_in_feats)
- for i in range(n_scales)])
-
- self.out_cumulative = out_cumulative
- self.return_only_hr = return_only_hr
-
- @property
- def num_scales(self):
- return len(self.heads)
-
- def forward(self, ms_inputs: List[torch.Tensor], smallest_scales_num: Optional[int] = None) \
- -> Union[torch.Tensor, List[torch.Tensor]]:
- """
- :param ms_inputs: List of inputs of different resolutions from HR to LR
- :param smallest_scales_num: int or None, number of smallest scales to take at input
- :return: Depending on return_only_hr:
- True: Only the most HR output
- False: List of outputs of different resolutions from HR to LR
- """
- if smallest_scales_num is None:
- assert len(self.heads) == len(ms_inputs), (len(self.heads), len(ms_inputs), smallest_scales_num)
- smallest_scales_num = len(self.heads)
- else:
- assert smallest_scales_num == len(ms_inputs) <= len(self.heads), (len(self.heads), len(ms_inputs), smallest_scales_num)
-
- cur_heads = self.heads[-smallest_scales_num:]
- ms_features = [cur_head(cur_inp) for cur_head, cur_inp in zip(cur_heads, ms_inputs)]
-
- all_outputs = []
- prev_tail_features = None
- for i in range(len(ms_features)):
- scale_i = -i - 1
-
- cur_tail_input = ms_features[-i - 1]
- if prev_tail_features is not None:
- if prev_tail_features.shape != cur_tail_input.shape:
- prev_tail_features = F.interpolate(prev_tail_features, size=cur_tail_input.shape[2:],
- mode='bilinear', align_corners=False)
- cur_tail_input = torch.cat((cur_tail_input, prev_tail_features), dim=1)
-
- cur_out, cur_tail_feats = self.tails[scale_i](cur_tail_input, return_last_act=True)
-
- prev_tail_features = cur_tail_feats
- all_outputs.append(cur_out)
-
- if self.out_cumulative:
- all_outputs_cum = [all_outputs[0]]
- for i in range(1, len(ms_features)):
- cur_out = all_outputs[i]
- cur_out_cum = cur_out + F.interpolate(all_outputs_cum[-1], size=cur_out.shape[2:],
- mode='bilinear', align_corners=False)
- all_outputs_cum.append(cur_out_cum)
- all_outputs = all_outputs_cum
-
- if self.return_only_hr:
- return all_outputs[-1]
- else:
- return all_outputs[::-1]
-
-
-class MultiscaleDiscriminatorSimple(nn.Module):
- def __init__(self, ms_impl):
- super().__init__()
- self.ms_impl = nn.ModuleList(ms_impl)
-
- @property
- def num_scales(self):
- return len(self.ms_impl)
-
- def forward(self, ms_inputs: List[torch.Tensor], smallest_scales_num: Optional[int] = None) \
- -> List[Tuple[torch.Tensor, List[torch.Tensor]]]:
- """
- :param ms_inputs: List of inputs of different resolutions from HR to LR
- :param smallest_scales_num: int or None, number of smallest scales to take at input
- :return: List of pairs (prediction, features) for different resolutions from HR to LR
- """
- if smallest_scales_num is None:
- assert len(self.ms_impl) == len(ms_inputs), (len(self.ms_impl), len(ms_inputs), smallest_scales_num)
- smallest_scales_num = len(self.heads)
- else:
- assert smallest_scales_num == len(ms_inputs) <= len(self.ms_impl), \
- (len(self.ms_impl), len(ms_inputs), smallest_scales_num)
-
- return [cur_discr(cur_input) for cur_discr, cur_input in zip(self.ms_impl[-smallest_scales_num:], ms_inputs)]
-
-
-class SingleToMultiScaleInputMixin:
- def forward(self, x: torch.Tensor) -> List:
- orig_height, orig_width = x.shape[2:]
- factors = [2 ** i for i in range(self.num_scales)]
- ms_inputs = [F.interpolate(x, size=(orig_height // f, orig_width // f), mode='bilinear', align_corners=False)
- for f in factors]
- return super().forward(ms_inputs)
-
-
-class GeneratorMultiToSingleOutputMixin:
- def forward(self, x):
- return super().forward(x)[0]
-
-
-class DiscriminatorMultiToSingleOutputMixin:
- def forward(self, x):
- out_feat_tuples = super().forward(x)
- return out_feat_tuples[0][0], [f for _, flist in out_feat_tuples for f in flist]
-
-
-class DiscriminatorMultiToSingleOutputStackedMixin:
- def __init__(self, *args, return_feats_only_levels=None, **kwargs):
- super().__init__(*args, **kwargs)
- self.return_feats_only_levels = return_feats_only_levels
-
- def forward(self, x):
- out_feat_tuples = super().forward(x)
- outs = [out for out, _ in out_feat_tuples]
- scaled_outs = [outs[0]] + [F.interpolate(cur_out, size=outs[0].shape[-2:],
- mode='bilinear', align_corners=False)
- for cur_out in outs[1:]]
- out = torch.cat(scaled_outs, dim=1)
- if self.return_feats_only_levels is not None:
- feat_lists = [out_feat_tuples[i][1] for i in self.return_feats_only_levels]
- else:
- feat_lists = [flist for _, flist in out_feat_tuples]
- feats = [f for flist in feat_lists for f in flist]
- return out, feats
-
-
-class MultiscaleDiscrSingleInput(SingleToMultiScaleInputMixin, DiscriminatorMultiToSingleOutputStackedMixin, MultiscaleDiscriminatorSimple):
- pass
-
-
-class MultiscaleResNetSingle(GeneratorMultiToSingleOutputMixin, SingleToMultiScaleInputMixin, MultiscaleResNet):
- pass
diff --git a/spaces/Jamel887/Rvc-tio887/lib/infer_pack/attentions.py b/spaces/Jamel887/Rvc-tio887/lib/infer_pack/attentions.py
deleted file mode 100644
index 05501be1871643f78dddbeaa529c96667031a8db..0000000000000000000000000000000000000000
--- a/spaces/Jamel887/Rvc-tio887/lib/infer_pack/attentions.py
+++ /dev/null
@@ -1,417 +0,0 @@
-import copy
-import math
-import numpy as np
-import torch
-from torch import nn
-from torch.nn import functional as F
-
-from lib.infer_pack import commons
-from lib.infer_pack import modules
-from lib.infer_pack.modules import LayerNorm
-
-
-class Encoder(nn.Module):
- def __init__(
- self,
- hidden_channels,
- filter_channels,
- n_heads,
- n_layers,
- kernel_size=1,
- p_dropout=0.0,
- window_size=10,
- **kwargs
- ):
- super().__init__()
- self.hidden_channels = hidden_channels
- self.filter_channels = filter_channels
- self.n_heads = n_heads
- self.n_layers = n_layers
- self.kernel_size = kernel_size
- self.p_dropout = p_dropout
- self.window_size = window_size
-
- self.drop = nn.Dropout(p_dropout)
- self.attn_layers = nn.ModuleList()
- self.norm_layers_1 = nn.ModuleList()
- self.ffn_layers = nn.ModuleList()
- self.norm_layers_2 = nn.ModuleList()
- for i in range(self.n_layers):
- self.attn_layers.append(
- MultiHeadAttention(
- hidden_channels,
- hidden_channels,
- n_heads,
- p_dropout=p_dropout,
- window_size=window_size,
- )
- )
- self.norm_layers_1.append(LayerNorm(hidden_channels))
- self.ffn_layers.append(
- FFN(
- hidden_channels,
- hidden_channels,
- filter_channels,
- kernel_size,
- p_dropout=p_dropout,
- )
- )
- self.norm_layers_2.append(LayerNorm(hidden_channels))
-
- def forward(self, x, x_mask):
- attn_mask = x_mask.unsqueeze(2) * x_mask.unsqueeze(-1)
- x = x * x_mask
- for i in range(self.n_layers):
- y = self.attn_layers[i](x, x, attn_mask)
- y = self.drop(y)
- x = self.norm_layers_1[i](x + y)
-
- y = self.ffn_layers[i](x, x_mask)
- y = self.drop(y)
- x = self.norm_layers_2[i](x + y)
- x = x * x_mask
- return x
-
-
-class Decoder(nn.Module):
- def __init__(
- self,
- hidden_channels,
- filter_channels,
- n_heads,
- n_layers,
- kernel_size=1,
- p_dropout=0.0,
- proximal_bias=False,
- proximal_init=True,
- **kwargs
- ):
- super().__init__()
- self.hidden_channels = hidden_channels
- self.filter_channels = filter_channels
- self.n_heads = n_heads
- self.n_layers = n_layers
- self.kernel_size = kernel_size
- self.p_dropout = p_dropout
- self.proximal_bias = proximal_bias
- self.proximal_init = proximal_init
-
- self.drop = nn.Dropout(p_dropout)
- self.self_attn_layers = nn.ModuleList()
- self.norm_layers_0 = nn.ModuleList()
- self.encdec_attn_layers = nn.ModuleList()
- self.norm_layers_1 = nn.ModuleList()
- self.ffn_layers = nn.ModuleList()
- self.norm_layers_2 = nn.ModuleList()
- for i in range(self.n_layers):
- self.self_attn_layers.append(
- MultiHeadAttention(
- hidden_channels,
- hidden_channels,
- n_heads,
- p_dropout=p_dropout,
- proximal_bias=proximal_bias,
- proximal_init=proximal_init,
- )
- )
- self.norm_layers_0.append(LayerNorm(hidden_channels))
- self.encdec_attn_layers.append(
- MultiHeadAttention(
- hidden_channels, hidden_channels, n_heads, p_dropout=p_dropout
- )
- )
- self.norm_layers_1.append(LayerNorm(hidden_channels))
- self.ffn_layers.append(
- FFN(
- hidden_channels,
- hidden_channels,
- filter_channels,
- kernel_size,
- p_dropout=p_dropout,
- causal=True,
- )
- )
- self.norm_layers_2.append(LayerNorm(hidden_channels))
-
- def forward(self, x, x_mask, h, h_mask):
- """
- x: decoder input
- h: encoder output
- """
- self_attn_mask = commons.subsequent_mask(x_mask.size(2)).to(
- device=x.device, dtype=x.dtype
- )
- encdec_attn_mask = h_mask.unsqueeze(2) * x_mask.unsqueeze(-1)
- x = x * x_mask
- for i in range(self.n_layers):
- y = self.self_attn_layers[i](x, x, self_attn_mask)
- y = self.drop(y)
- x = self.norm_layers_0[i](x + y)
-
- y = self.encdec_attn_layers[i](x, h, encdec_attn_mask)
- y = self.drop(y)
- x = self.norm_layers_1[i](x + y)
-
- y = self.ffn_layers[i](x, x_mask)
- y = self.drop(y)
- x = self.norm_layers_2[i](x + y)
- x = x * x_mask
- return x
-
-
-class MultiHeadAttention(nn.Module):
- def __init__(
- self,
- channels,
- out_channels,
- n_heads,
- p_dropout=0.0,
- window_size=None,
- heads_share=True,
- block_length=None,
- proximal_bias=False,
- proximal_init=False,
- ):
- super().__init__()
- assert channels % n_heads == 0
-
- self.channels = channels
- self.out_channels = out_channels
- self.n_heads = n_heads
- self.p_dropout = p_dropout
- self.window_size = window_size
- self.heads_share = heads_share
- self.block_length = block_length
- self.proximal_bias = proximal_bias
- self.proximal_init = proximal_init
- self.attn = None
-
- self.k_channels = channels // n_heads
- self.conv_q = nn.Conv1d(channels, channels, 1)
- self.conv_k = nn.Conv1d(channels, channels, 1)
- self.conv_v = nn.Conv1d(channels, channels, 1)
- self.conv_o = nn.Conv1d(channels, out_channels, 1)
- self.drop = nn.Dropout(p_dropout)
-
- if window_size is not None:
- n_heads_rel = 1 if heads_share else n_heads
- rel_stddev = self.k_channels**-0.5
- self.emb_rel_k = nn.Parameter(
- torch.randn(n_heads_rel, window_size * 2 + 1, self.k_channels)
- * rel_stddev
- )
- self.emb_rel_v = nn.Parameter(
- torch.randn(n_heads_rel, window_size * 2 + 1, self.k_channels)
- * rel_stddev
- )
-
- nn.init.xavier_uniform_(self.conv_q.weight)
- nn.init.xavier_uniform_(self.conv_k.weight)
- nn.init.xavier_uniform_(self.conv_v.weight)
- if proximal_init:
- with torch.no_grad():
- self.conv_k.weight.copy_(self.conv_q.weight)
- self.conv_k.bias.copy_(self.conv_q.bias)
-
- def forward(self, x, c, attn_mask=None):
- q = self.conv_q(x)
- k = self.conv_k(c)
- v = self.conv_v(c)
-
- x, self.attn = self.attention(q, k, v, mask=attn_mask)
-
- x = self.conv_o(x)
- return x
-
- def attention(self, query, key, value, mask=None):
- # reshape [b, d, t] -> [b, n_h, t, d_k]
- b, d, t_s, t_t = (*key.size(), query.size(2))
- query = query.view(b, self.n_heads, self.k_channels, t_t).transpose(2, 3)
- key = key.view(b, self.n_heads, self.k_channels, t_s).transpose(2, 3)
- value = value.view(b, self.n_heads, self.k_channels, t_s).transpose(2, 3)
-
- scores = torch.matmul(query / math.sqrt(self.k_channels), key.transpose(-2, -1))
- if self.window_size is not None:
- assert (
- t_s == t_t
- ), "Relative attention is only available for self-attention."
- key_relative_embeddings = self._get_relative_embeddings(self.emb_rel_k, t_s)
- rel_logits = self._matmul_with_relative_keys(
- query / math.sqrt(self.k_channels), key_relative_embeddings
- )
- scores_local = self._relative_position_to_absolute_position(rel_logits)
- scores = scores + scores_local
- if self.proximal_bias:
- assert t_s == t_t, "Proximal bias is only available for self-attention."
- scores = scores + self._attention_bias_proximal(t_s).to(
- device=scores.device, dtype=scores.dtype
- )
- if mask is not None:
- scores = scores.masked_fill(mask == 0, -1e4)
- if self.block_length is not None:
- assert (
- t_s == t_t
- ), "Local attention is only available for self-attention."
- block_mask = (
- torch.ones_like(scores)
- .triu(-self.block_length)
- .tril(self.block_length)
- )
- scores = scores.masked_fill(block_mask == 0, -1e4)
- p_attn = F.softmax(scores, dim=-1) # [b, n_h, t_t, t_s]
- p_attn = self.drop(p_attn)
- output = torch.matmul(p_attn, value)
- if self.window_size is not None:
- relative_weights = self._absolute_position_to_relative_position(p_attn)
- value_relative_embeddings = self._get_relative_embeddings(
- self.emb_rel_v, t_s
- )
- output = output + self._matmul_with_relative_values(
- relative_weights, value_relative_embeddings
- )
- output = (
- output.transpose(2, 3).contiguous().view(b, d, t_t)
- ) # [b, n_h, t_t, d_k] -> [b, d, t_t]
- return output, p_attn
-
- def _matmul_with_relative_values(self, x, y):
- """
- x: [b, h, l, m]
- y: [h or 1, m, d]
- ret: [b, h, l, d]
- """
- ret = torch.matmul(x, y.unsqueeze(0))
- return ret
-
- def _matmul_with_relative_keys(self, x, y):
- """
- x: [b, h, l, d]
- y: [h or 1, m, d]
- ret: [b, h, l, m]
- """
- ret = torch.matmul(x, y.unsqueeze(0).transpose(-2, -1))
- return ret
-
- def _get_relative_embeddings(self, relative_embeddings, length):
- max_relative_position = 2 * self.window_size + 1
- # Pad first before slice to avoid using cond ops.
- pad_length = max(length - (self.window_size + 1), 0)
- slice_start_position = max((self.window_size + 1) - length, 0)
- slice_end_position = slice_start_position + 2 * length - 1
- if pad_length > 0:
- padded_relative_embeddings = F.pad(
- relative_embeddings,
- commons.convert_pad_shape([[0, 0], [pad_length, pad_length], [0, 0]]),
- )
- else:
- padded_relative_embeddings = relative_embeddings
- used_relative_embeddings = padded_relative_embeddings[
- :, slice_start_position:slice_end_position
- ]
- return used_relative_embeddings
-
- def _relative_position_to_absolute_position(self, x):
- """
- x: [b, h, l, 2*l-1]
- ret: [b, h, l, l]
- """
- batch, heads, length, _ = x.size()
- # Concat columns of pad to shift from relative to absolute indexing.
- x = F.pad(x, commons.convert_pad_shape([[0, 0], [0, 0], [0, 0], [0, 1]]))
-
- # Concat extra elements so to add up to shape (len+1, 2*len-1).
- x_flat = x.view([batch, heads, length * 2 * length])
- x_flat = F.pad(
- x_flat, commons.convert_pad_shape([[0, 0], [0, 0], [0, length - 1]])
- )
-
- # Reshape and slice out the padded elements.
- x_final = x_flat.view([batch, heads, length + 1, 2 * length - 1])[
- :, :, :length, length - 1 :
- ]
- return x_final
-
- def _absolute_position_to_relative_position(self, x):
- """
- x: [b, h, l, l]
- ret: [b, h, l, 2*l-1]
- """
- batch, heads, length, _ = x.size()
- # padd along column
- x = F.pad(
- x, commons.convert_pad_shape([[0, 0], [0, 0], [0, 0], [0, length - 1]])
- )
- x_flat = x.view([batch, heads, length**2 + length * (length - 1)])
- # add 0's in the beginning that will skew the elements after reshape
- x_flat = F.pad(x_flat, commons.convert_pad_shape([[0, 0], [0, 0], [length, 0]]))
- x_final = x_flat.view([batch, heads, length, 2 * length])[:, :, :, 1:]
- return x_final
-
- def _attention_bias_proximal(self, length):
- """Bias for self-attention to encourage attention to close positions.
- Args:
- length: an integer scalar.
- Returns:
- a Tensor with shape [1, 1, length, length]
- """
- r = torch.arange(length, dtype=torch.float32)
- diff = torch.unsqueeze(r, 0) - torch.unsqueeze(r, 1)
- return torch.unsqueeze(torch.unsqueeze(-torch.log1p(torch.abs(diff)), 0), 0)
-
-
-class FFN(nn.Module):
- def __init__(
- self,
- in_channels,
- out_channels,
- filter_channels,
- kernel_size,
- p_dropout=0.0,
- activation=None,
- causal=False,
- ):
- super().__init__()
- self.in_channels = in_channels
- self.out_channels = out_channels
- self.filter_channels = filter_channels
- self.kernel_size = kernel_size
- self.p_dropout = p_dropout
- self.activation = activation
- self.causal = causal
-
- if causal:
- self.padding = self._causal_padding
- else:
- self.padding = self._same_padding
-
- self.conv_1 = nn.Conv1d(in_channels, filter_channels, kernel_size)
- self.conv_2 = nn.Conv1d(filter_channels, out_channels, kernel_size)
- self.drop = nn.Dropout(p_dropout)
-
- def forward(self, x, x_mask):
- x = self.conv_1(self.padding(x * x_mask))
- if self.activation == "gelu":
- x = x * torch.sigmoid(1.702 * x)
- else:
- x = torch.relu(x)
- x = self.drop(x)
- x = self.conv_2(self.padding(x * x_mask))
- return x * x_mask
-
- def _causal_padding(self, x):
- if self.kernel_size == 1:
- return x
- pad_l = self.kernel_size - 1
- pad_r = 0
- padding = [[0, 0], [0, 0], [pad_l, pad_r]]
- x = F.pad(x, commons.convert_pad_shape(padding))
- return x
-
- def _same_padding(self, x):
- if self.kernel_size == 1:
- return x
- pad_l = (self.kernel_size - 1) // 2
- pad_r = self.kernel_size // 2
- padding = [[0, 0], [0, 0], [pad_l, pad_r]]
- x = F.pad(x, commons.convert_pad_shape(padding))
- return x
diff --git a/spaces/Jonni/04-Gradio_SOTA/app.py b/spaces/Jonni/04-Gradio_SOTA/app.py
deleted file mode 100644
index c1cd92499cf1c7d2a91b4dc226bf2d558ff67661..0000000000000000000000000000000000000000
--- a/spaces/Jonni/04-Gradio_SOTA/app.py
+++ /dev/null
@@ -1,51 +0,0 @@
-import gradio as gr
-from qasrl_model_pipeline import QASRL_Pipeline
-
-models = ["kleinay/qanom-seq2seq-model-baseline",
- "kleinay/qanom-seq2seq-model-joint"]
-pipelines = {model: QASRL_Pipeline(model) for model in models}
-
-
-description = f"""Using Seq2Seq T5 model which takes a sequence of items and outputs another sequence this model generates Questions and Answers (QA) with focus on Semantic Role Labeling (SRL)"""
-title="Seq2Seq T5 Questions and Answers (QA) with Semantic Role Labeling (SRL)"
-examples = [[models[0], "In March and April the patient
had two falls. One was related to asthma, heart palpitations. The second was due to syncope and post covid vaccination dizziness during exercise. The patient is now getting an EKG. Former EKG had shown that there was a bundle branch block. Patient had some uncontrolled immune system reactions like anaphylaxis and shortness of breath.", True, "fall"],
- [models[1], "In March and April the patient had two falls. One was related to asthma, heart palpitations. The second was due to syncope and post covid vaccination dizziness during exercise. The patient is now getting an EKG. Former EKG had shown that there was a bundle branch block. Patient had some uncontrolled immune system reactions
like anaphylaxis and shortness of breath.", True, "reactions"],
- [models[0], "In March and April the patient had two falls. One was related
to asthma, heart palpitations. The second was due to syncope and post covid vaccination dizziness during exercise. The patient is now getting an EKG. Former EKG had shown that there was a bundle branch block. Patient had some uncontrolled immune system reactions like anaphylaxis and shortness of breath.", True, "relate"],
- [models[1], "In March and April the patient
had two falls. One was related to asthma, heart palpitations. The second was due to syncope and post covid vaccination dizziness during exercise. The patient is now getting an EKG. Former EKG had shown that there was a bundle branch block. Patient had some uncontrolled immune system reactions like anaphylaxis and shortness of breath.", False, "fall"]]
-
-input_sent_box_label = "Insert sentence here. Mark the predicate by adding the token '
' before it."
-verb_form_inp_placeholder = "e.g. 'decide' for the nominalization 'decision', 'teach' for 'teacher', etc."
-links = """
'.format(t)
- for t in torch_versions
- ]
- )
-
-
-if __name__ == "__main__":
- parser = argparse.ArgumentParser()
- parser.add_argument("--d2-version", help="detectron2 version number, default to empty")
- args = parser.parse_args()
- d2_version = f"=={args.d2_version}" if args.d2_version else ""
-
- all_versions = (
- [("1.8", k) for k in ["11.1", "10.2", "10.1", "cpu"]]
- + [("1.9", k) for k in ["11.1", "10.2", "cpu"]]
- + [("1.10", k) for k in ["11.3", "11.1", "10.2", "cpu"]]
- )
-
- torch_versions = sorted(
- {k[0] for k in all_versions}, key=lambda x: int(x.split(".")[1]), reverse=True
- )
- cuda_versions = sorted(
- {k[1] for k in all_versions}, key=lambda x: float(x) if x != "cpu" else 0, reverse=True
- )
-
- table = gen_header(torch_versions)
- for cu in cuda_versions:
- table += f"""
{cu}
"""
- cu_suffix = CUDA_SUFFIX[cu]
- for torch in torch_versions:
- if (torch, cu) in all_versions:
- cell = template.format(d2_version=d2_version, cuda=cu_suffix, torch=torch)
- else:
- cell = ""
- table += f"""
{cell}
"""
- table += "
"
- table += "
"
- print(table)
diff --git a/spaces/OpenGVLab/InternGPT/third-party/lama/bin/models/ade20k/utils.py b/spaces/OpenGVLab/InternGPT/third-party/lama/bin/models/ade20k/utils.py
deleted file mode 100644
index f337db7db54c82be041698d694e1403e8918c4c0..0000000000000000000000000000000000000000
--- a/spaces/OpenGVLab/InternGPT/third-party/lama/bin/models/ade20k/utils.py
+++ /dev/null
@@ -1,40 +0,0 @@
-"""Modified from https://github.com/CSAILVision/semantic-segmentation-pytorch"""
-
-import os
-import sys
-
-import numpy as np
-import torch
-
-try:
- from urllib import urlretrieve
-except ImportError:
- from urllib.request import urlretrieve
-
-
-def load_url(url, model_dir='./pretrained', map_location=None):
- if not os.path.exists(model_dir):
- os.makedirs(model_dir)
- filename = url.split('/')[-1]
- cached_file = os.path.join(model_dir, filename)
- if not os.path.exists(cached_file):
- sys.stderr.write('Downloading: "{}" to {}\n'.format(url, cached_file))
- urlretrieve(url, cached_file)
- return torch.load(cached_file, map_location=map_location)
-
-
-def color_encode(labelmap, colors, mode='RGB'):
- labelmap = labelmap.astype('int')
- labelmap_rgb = np.zeros((labelmap.shape[0], labelmap.shape[1], 3),
- dtype=np.uint8)
- for label in np.unique(labelmap):
- if label < 0:
- continue
- labelmap_rgb += (labelmap == label)[:, :, np.newaxis] * \
- np.tile(colors[label],
- (labelmap.shape[0], labelmap.shape[1], 1))
-
- if mode == 'BGR':
- return labelmap_rgb[:, :, ::-1]
- else:
- return labelmap_rgb
diff --git a/spaces/PAIR/Text2Video-Zero/annotator/uniformer/mmseg/models/backbones/vit.py b/spaces/PAIR/Text2Video-Zero/annotator/uniformer/mmseg/models/backbones/vit.py
deleted file mode 100644
index 59e4479650690e08cbc4cab9427aefda47c2116d..0000000000000000000000000000000000000000
--- a/spaces/PAIR/Text2Video-Zero/annotator/uniformer/mmseg/models/backbones/vit.py
+++ /dev/null
@@ -1,459 +0,0 @@
-"""Modified from https://github.com/rwightman/pytorch-image-
-models/blob/master/timm/models/vision_transformer.py."""
-
-import math
-
-import torch
-import torch.nn as nn
-import torch.nn.functional as F
-import torch.utils.checkpoint as cp
-from annotator.uniformer.mmcv.cnn import (Conv2d, Linear, build_activation_layer, build_norm_layer,
- constant_init, kaiming_init, normal_init)
-from annotator.uniformer.mmcv.runner import _load_checkpoint
-from annotator.uniformer.mmcv.utils.parrots_wrapper import _BatchNorm
-
-from annotator.uniformer.mmseg.utils import get_root_logger
-from ..builder import BACKBONES
-from ..utils import DropPath, trunc_normal_
-
-
-class Mlp(nn.Module):
- """MLP layer for Encoder block.
-
- Args:
- in_features(int): Input dimension for the first fully
- connected layer.
- hidden_features(int): Output dimension for the first fully
- connected layer.
- out_features(int): Output dementsion for the second fully
- connected layer.
- act_cfg(dict): Config dict for activation layer.
- Default: dict(type='GELU').
- drop(float): Drop rate for the dropout layer. Dropout rate has
- to be between 0 and 1. Default: 0.
- """
-
- def __init__(self,
- in_features,
- hidden_features=None,
- out_features=None,
- act_cfg=dict(type='GELU'),
- drop=0.):
- super(Mlp, self).__init__()
- out_features = out_features or in_features
- hidden_features = hidden_features or in_features
- self.fc1 = Linear(in_features, hidden_features)
- self.act = build_activation_layer(act_cfg)
- self.fc2 = Linear(hidden_features, out_features)
- self.drop = nn.Dropout(drop)
-
- def forward(self, x):
- x = self.fc1(x)
- x = self.act(x)
- x = self.drop(x)
- x = self.fc2(x)
- x = self.drop(x)
- return x
-
-
-class Attention(nn.Module):
- """Attention layer for Encoder block.
-
- Args:
- dim (int): Dimension for the input vector.
- num_heads (int): Number of parallel attention heads.
- qkv_bias (bool): Enable bias for qkv if True. Default: False.
- qk_scale (float): Override default qk scale of head_dim ** -0.5 if set.
- attn_drop (float): Drop rate for attention output weights.
- Default: 0.
- proj_drop (float): Drop rate for output weights. Default: 0.
- """
-
- def __init__(self,
- dim,
- num_heads=8,
- qkv_bias=False,
- qk_scale=None,
- attn_drop=0.,
- proj_drop=0.):
- super(Attention, self).__init__()
- self.num_heads = num_heads
- head_dim = dim // num_heads
- self.scale = qk_scale or head_dim**-0.5
-
- self.qkv = nn.Linear(dim, dim * 3, bias=qkv_bias)
- self.attn_drop = nn.Dropout(attn_drop)
- self.proj = Linear(dim, dim)
- self.proj_drop = nn.Dropout(proj_drop)
-
- def forward(self, x):
- b, n, c = x.shape
- qkv = self.qkv(x).reshape(b, n, 3, self.num_heads,
- c // self.num_heads).permute(2, 0, 3, 1, 4)
- q, k, v = qkv[0], qkv[1], qkv[2]
-
- attn = (q @ k.transpose(-2, -1)) * self.scale
- attn = attn.softmax(dim=-1)
- attn = self.attn_drop(attn)
-
- x = (attn @ v).transpose(1, 2).reshape(b, n, c)
- x = self.proj(x)
- x = self.proj_drop(x)
- return x
-
-
-class Block(nn.Module):
- """Implements encoder block with residual connection.
-
- Args:
- dim (int): The feature dimension.
- num_heads (int): Number of parallel attention heads.
- mlp_ratio (int): Ratio of mlp hidden dim to embedding dim.
- qk_scale (float): Override default qk scale of head_dim ** -0.5 if set.
- drop (float): Drop rate for mlp output weights. Default: 0.
- attn_drop (float): Drop rate for attention output weights.
- Default: 0.
- proj_drop (float): Drop rate for attn layer output weights.
- Default: 0.
- drop_path (float): Drop rate for paths of model.
- Default: 0.
- act_cfg (dict): Config dict for activation layer.
- Default: dict(type='GELU').
- norm_cfg (dict): Config dict for normalization layer.
- Default: dict(type='LN', requires_grad=True).
- with_cp (bool): Use checkpoint or not. Using checkpoint will save some
- memory while slowing down the training speed. Default: False.
- """
-
- def __init__(self,
- dim,
- num_heads,
- mlp_ratio=4,
- qkv_bias=False,
- qk_scale=None,
- drop=0.,
- attn_drop=0.,
- proj_drop=0.,
- drop_path=0.,
- act_cfg=dict(type='GELU'),
- norm_cfg=dict(type='LN', eps=1e-6),
- with_cp=False):
- super(Block, self).__init__()
- self.with_cp = with_cp
- _, self.norm1 = build_norm_layer(norm_cfg, dim)
- self.attn = Attention(dim, num_heads, qkv_bias, qk_scale, attn_drop,
- proj_drop)
- self.drop_path = DropPath(
- drop_path) if drop_path > 0. else nn.Identity()
- _, self.norm2 = build_norm_layer(norm_cfg, dim)
- mlp_hidden_dim = int(dim * mlp_ratio)
- self.mlp = Mlp(
- in_features=dim,
- hidden_features=mlp_hidden_dim,
- act_cfg=act_cfg,
- drop=drop)
-
- def forward(self, x):
-
- def _inner_forward(x):
- out = x + self.drop_path(self.attn(self.norm1(x)))
- out = out + self.drop_path(self.mlp(self.norm2(out)))
- return out
-
- if self.with_cp and x.requires_grad:
- out = cp.checkpoint(_inner_forward, x)
- else:
- out = _inner_forward(x)
-
- return out
-
-
-class PatchEmbed(nn.Module):
- """Image to Patch Embedding.
-
- Args:
- img_size (int | tuple): Input image size.
- default: 224.
- patch_size (int): Width and height for a patch.
- default: 16.
- in_channels (int): Input channels for images. Default: 3.
- embed_dim (int): The embedding dimension. Default: 768.
- """
-
- def __init__(self,
- img_size=224,
- patch_size=16,
- in_channels=3,
- embed_dim=768):
- super(PatchEmbed, self).__init__()
- if isinstance(img_size, int):
- self.img_size = (img_size, img_size)
- elif isinstance(img_size, tuple):
- self.img_size = img_size
- else:
- raise TypeError('img_size must be type of int or tuple')
- h, w = self.img_size
- self.patch_size = (patch_size, patch_size)
- self.num_patches = (h // patch_size) * (w // patch_size)
- self.proj = Conv2d(
- in_channels, embed_dim, kernel_size=patch_size, stride=patch_size)
-
- def forward(self, x):
- return self.proj(x).flatten(2).transpose(1, 2)
-
-
-@BACKBONES.register_module()
-class VisionTransformer(nn.Module):
- """Vision transformer backbone.
-
- A PyTorch impl of : `An Image is Worth 16x16 Words: Transformers for
- Image Recognition at Scale` - https://arxiv.org/abs/2010.11929
-
- Args:
- img_size (tuple): input image size. Default: (224, 224).
- patch_size (int, tuple): patch size. Default: 16.
- in_channels (int): number of input channels. Default: 3.
- embed_dim (int): embedding dimension. Default: 768.
- depth (int): depth of transformer. Default: 12.
- num_heads (int): number of attention heads. Default: 12.
- mlp_ratio (int): ratio of mlp hidden dim to embedding dim.
- Default: 4.
- out_indices (list | tuple | int): Output from which stages.
- Default: -1.
- qkv_bias (bool): enable bias for qkv if True. Default: True.
- qk_scale (float): override default qk scale of head_dim ** -0.5 if set.
- drop_rate (float): dropout rate. Default: 0.
- attn_drop_rate (float): attention dropout rate. Default: 0.
- drop_path_rate (float): Rate of DropPath. Default: 0.
- norm_cfg (dict): Config dict for normalization layer.
- Default: dict(type='LN', eps=1e-6, requires_grad=True).
- act_cfg (dict): Config dict for activation layer.
- Default: dict(type='GELU').
- norm_eval (bool): Whether to set norm layers to eval mode, namely,
- freeze running stats (mean and var). Note: Effect on Batch Norm
- and its variants only. Default: False.
- final_norm (bool): Whether to add a additional layer to normalize
- final feature map. Default: False.
- interpolate_mode (str): Select the interpolate mode for position
- embeding vector resize. Default: bicubic.
- with_cls_token (bool): If concatenating class token into image tokens
- as transformer input. Default: True.
- with_cp (bool): Use checkpoint or not. Using checkpoint
- will save some memory while slowing down the training speed.
- Default: False.
- """
-
- def __init__(self,
- img_size=(224, 224),
- patch_size=16,
- in_channels=3,
- embed_dim=768,
- depth=12,
- num_heads=12,
- mlp_ratio=4,
- out_indices=11,
- qkv_bias=True,
- qk_scale=None,
- drop_rate=0.,
- attn_drop_rate=0.,
- drop_path_rate=0.,
- norm_cfg=dict(type='LN', eps=1e-6, requires_grad=True),
- act_cfg=dict(type='GELU'),
- norm_eval=False,
- final_norm=False,
- with_cls_token=True,
- interpolate_mode='bicubic',
- with_cp=False):
- super(VisionTransformer, self).__init__()
- self.img_size = img_size
- self.patch_size = patch_size
- self.features = self.embed_dim = embed_dim
- self.patch_embed = PatchEmbed(
- img_size=img_size,
- patch_size=patch_size,
- in_channels=in_channels,
- embed_dim=embed_dim)
-
- self.with_cls_token = with_cls_token
- self.cls_token = nn.Parameter(torch.zeros(1, 1, self.embed_dim))
- self.pos_embed = nn.Parameter(
- torch.zeros(1, self.patch_embed.num_patches + 1, embed_dim))
- self.pos_drop = nn.Dropout(p=drop_rate)
-
- if isinstance(out_indices, int):
- self.out_indices = [out_indices]
- elif isinstance(out_indices, list) or isinstance(out_indices, tuple):
- self.out_indices = out_indices
- else:
- raise TypeError('out_indices must be type of int, list or tuple')
-
- dpr = [x.item() for x in torch.linspace(0, drop_path_rate, depth)
- ] # stochastic depth decay rule
- self.blocks = nn.ModuleList([
- Block(
- dim=embed_dim,
- num_heads=num_heads,
- mlp_ratio=mlp_ratio,
- qkv_bias=qkv_bias,
- qk_scale=qk_scale,
- drop=dpr[i],
- attn_drop=attn_drop_rate,
- act_cfg=act_cfg,
- norm_cfg=norm_cfg,
- with_cp=with_cp) for i in range(depth)
- ])
-
- self.interpolate_mode = interpolate_mode
- self.final_norm = final_norm
- if final_norm:
- _, self.norm = build_norm_layer(norm_cfg, embed_dim)
-
- self.norm_eval = norm_eval
- self.with_cp = with_cp
-
- def init_weights(self, pretrained=None):
- if isinstance(pretrained, str):
- logger = get_root_logger()
- checkpoint = _load_checkpoint(pretrained, logger=logger)
- if 'state_dict' in checkpoint:
- state_dict = checkpoint['state_dict']
- else:
- state_dict = checkpoint
-
- if 'pos_embed' in state_dict.keys():
- if self.pos_embed.shape != state_dict['pos_embed'].shape:
- logger.info(msg=f'Resize the pos_embed shape from \
-{state_dict["pos_embed"].shape} to {self.pos_embed.shape}')
- h, w = self.img_size
- pos_size = int(
- math.sqrt(state_dict['pos_embed'].shape[1] - 1))
- state_dict['pos_embed'] = self.resize_pos_embed(
- state_dict['pos_embed'], (h, w), (pos_size, pos_size),
- self.patch_size, self.interpolate_mode)
-
- self.load_state_dict(state_dict, False)
-
- elif pretrained is None:
- # We only implement the 'jax_impl' initialization implemented at
- # https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py#L353 # noqa: E501
- trunc_normal_(self.pos_embed, std=.02)
- trunc_normal_(self.cls_token, std=.02)
- for n, m in self.named_modules():
- if isinstance(m, Linear):
- trunc_normal_(m.weight, std=.02)
- if m.bias is not None:
- if 'mlp' in n:
- normal_init(m.bias, std=1e-6)
- else:
- constant_init(m.bias, 0)
- elif isinstance(m, Conv2d):
- kaiming_init(m.weight, mode='fan_in')
- if m.bias is not None:
- constant_init(m.bias, 0)
- elif isinstance(m, (_BatchNorm, nn.GroupNorm, nn.LayerNorm)):
- constant_init(m.bias, 0)
- constant_init(m.weight, 1.0)
- else:
- raise TypeError('pretrained must be a str or None')
-
- def _pos_embeding(self, img, patched_img, pos_embed):
- """Positiong embeding method.
-
- Resize the pos_embed, if the input image size doesn't match
- the training size.
- Args:
- img (torch.Tensor): The inference image tensor, the shape
- must be [B, C, H, W].
- patched_img (torch.Tensor): The patched image, it should be
- shape of [B, L1, C].
- pos_embed (torch.Tensor): The pos_embed weighs, it should be
- shape of [B, L2, c].
- Return:
- torch.Tensor: The pos encoded image feature.
- """
- assert patched_img.ndim == 3 and pos_embed.ndim == 3, \
- 'the shapes of patched_img and pos_embed must be [B, L, C]'
- x_len, pos_len = patched_img.shape[1], pos_embed.shape[1]
- if x_len != pos_len:
- if pos_len == (self.img_size[0] // self.patch_size) * (
- self.img_size[1] // self.patch_size) + 1:
- pos_h = self.img_size[0] // self.patch_size
- pos_w = self.img_size[1] // self.patch_size
- else:
- raise ValueError(
- 'Unexpected shape of pos_embed, got {}.'.format(
- pos_embed.shape))
- pos_embed = self.resize_pos_embed(pos_embed, img.shape[2:],
- (pos_h, pos_w), self.patch_size,
- self.interpolate_mode)
- return self.pos_drop(patched_img + pos_embed)
-
- @staticmethod
- def resize_pos_embed(pos_embed, input_shpae, pos_shape, patch_size, mode):
- """Resize pos_embed weights.
-
- Resize pos_embed using bicubic interpolate method.
- Args:
- pos_embed (torch.Tensor): pos_embed weights.
- input_shpae (tuple): Tuple for (input_h, intput_w).
- pos_shape (tuple): Tuple for (pos_h, pos_w).
- patch_size (int): Patch size.
- Return:
- torch.Tensor: The resized pos_embed of shape [B, L_new, C]
- """
- assert pos_embed.ndim == 3, 'shape of pos_embed must be [B, L, C]'
- input_h, input_w = input_shpae
- pos_h, pos_w = pos_shape
- cls_token_weight = pos_embed[:, 0]
- pos_embed_weight = pos_embed[:, (-1 * pos_h * pos_w):]
- pos_embed_weight = pos_embed_weight.reshape(
- 1, pos_h, pos_w, pos_embed.shape[2]).permute(0, 3, 1, 2)
- pos_embed_weight = F.interpolate(
- pos_embed_weight,
- size=[input_h // patch_size, input_w // patch_size],
- align_corners=False,
- mode=mode)
- cls_token_weight = cls_token_weight.unsqueeze(1)
- pos_embed_weight = torch.flatten(pos_embed_weight, 2).transpose(1, 2)
- pos_embed = torch.cat((cls_token_weight, pos_embed_weight), dim=1)
- return pos_embed
-
- def forward(self, inputs):
- B = inputs.shape[0]
-
- x = self.patch_embed(inputs)
-
- cls_tokens = self.cls_token.expand(B, -1, -1)
- x = torch.cat((cls_tokens, x), dim=1)
- x = self._pos_embeding(inputs, x, self.pos_embed)
-
- if not self.with_cls_token:
- # Remove class token for transformer input
- x = x[:, 1:]
-
- outs = []
- for i, blk in enumerate(self.blocks):
- x = blk(x)
- if i == len(self.blocks) - 1:
- if self.final_norm:
- x = self.norm(x)
- if i in self.out_indices:
- if self.with_cls_token:
- # Remove class token and reshape token for decoder head
- out = x[:, 1:]
- else:
- out = x
- B, _, C = out.shape
- out = out.reshape(B, inputs.shape[2] // self.patch_size,
- inputs.shape[3] // self.patch_size,
- C).permute(0, 3, 1, 2)
- outs.append(out)
-
- return tuple(outs)
-
- def train(self, mode=True):
- super(VisionTransformer, self).train(mode)
- if mode and self.norm_eval:
- for m in self.modules():
- if isinstance(m, nn.LayerNorm):
- m.eval()
diff --git a/spaces/Pattr/DrumClassification/lilypond-2.24.2/lib/guile/2.2/ccache/ice-9/mapping.go b/spaces/Pattr/DrumClassification/lilypond-2.24.2/lib/guile/2.2/ccache/ice-9/mapping.go
deleted file mode 100644
index 75c73d9f010c2a2ee974593f75208232d12e8ad0..0000000000000000000000000000000000000000
Binary files a/spaces/Pattr/DrumClassification/lilypond-2.24.2/lib/guile/2.2/ccache/ice-9/mapping.go and /dev/null differ
diff --git a/spaces/PaulEdwards/StarWords/app.py b/spaces/PaulEdwards/StarWords/app.py
deleted file mode 100644
index ba04fd113139f6ad2a8708e00ef5027be3941cd7..0000000000000000000000000000000000000000
--- a/spaces/PaulEdwards/StarWords/app.py
+++ /dev/null
@@ -1,1036 +0,0 @@
-import gradio as gr
-from transformers import pipeline
-title = "Starwars words..."
-examples = [
- ["Did you hear that? They've shut down the main reactor. We'll be destroyed for sure. This is madness!"],
- ["We're doomed!"],
- ["There'll be no escape for the Princess this time."],
- ["What's that?"],
- ["I should have known better than to trust the logic of a half-sized thermocapsulary dehousing assister..."],
- ["Hurry up! Come with me! What are you waiting for?! Get in gear!"],
- ["Artoo! Artoo-Detoo, where are you?"],
- ["At last! Where have you been?"],
- ["They're heading in this direction. What are we going to do? We'll be sent to the spice mines of Kessel or smashed into who knows what!"],
- ["Wait a minute, where are you going?"],
- ["The Death Star plans are not in the main computer."],
- ["Where are those transmissions you intercepted?"],
- ["We intercepted no transmissions. Aaah... This is a consular ship. Were on a diplomatic mission."],
- ["If this is a consular ship... where is the Ambassador?"],
- ["Commander, tear this ship apart until you've found those plans and bring me the Ambassador. I want her alive!"],
- ["There she is! Set for stun!"],
- ["She'll be all right. Inform Lord Vader we have a prisoner."],
- ["Hey, you're not permitted in there. It's restricted. You'll be deactivated for sure.."],
- ["Don't call me a mindless philosopher, you overweight glob of grease! Now come out before somebody sees you."],
- ["Secret mission? What plans? What are you talking about? I'm not getting in there!"],
- ["I'm going to regret this."],
- ["There goes another one."],
- ["Hold your fire. There are no life forms. It must have been short-circuited."],
- ["That's funny, the damage doesn't look as bad from out here."],
- ["Are you sure this things safe?"],
- ["I've told you kids to slow down!"],
- ["Did I hear a young noise blast through here?"],
- ["It was just Wormie on another rampage."],
- ["Shape it up you guys!... Biggs?"],
- ["I didn't know you were back! When did you get in?"],
- ["Just now. I wanted to surprise you, hot shot. I thought you'd be here... certainly didn't expect you to be out working. "],
- ["The Academy didn't change you much... but you're back so soon? Hey, what happened, didn't you get your commission?"],
- ["Of course I got it. Signed aboard The Rand Ecliptic last week. First mate Biggs Darklighter at your service...... I just came back to say goodbye to all you unfortunate landlocked simpletons."],
- ["I almost forgot. There's a battle going on! Right here in our system. Come and look!"],
- ["Not again! Forget it."],
- ["There they are!"],
- ["That's no battle, hot shot... they're just sitting there! Probably a freighter-tanker refueling."],
- ["But there was a lot of firing earlier..."],
- ["Hey, easy with those..."],
- ["Don't worry about it, Wormie."],
- ["I keep telling you, the Rebellion is a long way from here. I doubt if the Empire would even fight to keep this system. Believe me Luke, this planet is a big hunk of nothing..."],
- ["Lord Vader, I should have known. Only you could be so bold. The Imperial Senate will not sit stillfor this, when they hear you've attacked a diplomatic..."],
- ["Don't play games with me, Your Highness. You weren't on any mercy mission this time. You passed directly through a restricted system. Several transmissions were beamed to this ship by Rebel spies. I want to know what happened to the plans they sent you."],
- ["I don't know what you're talking about. I'm a member of the Imperial Senate on a diplomatic mission to Alderaan..."],
- ["You're a part of the Rebel Alliance... and a traitor. Take her away!"],
- ["Holding her is dangerous. If word of this gets out, it could generate sympathy for the Rebellion in the senate."],
- ["I have traced the Rebel spies to her. Now she is my only link to find their secret base!"],
- ["you anything."],
- ["Leave that to me. Send a distress signal and then inform the senate that all aboard were killed!"],
- ["Lord Vader, the battle station plans are not aboard this ship! And no transmissions were made. An escape pod was jettisoned during the fighting, but no life forms were aboard."],
- ["She must have hidden the plans in the escape pod. Send a detachment down to retrieve them. See to it personally, Commander. There'll be no one to stop us this time."],
- ["Yes, sir."],
- ["How did I get into this mess? I really don't know how. We seem to be made to suffer. It's our lot in life."],
- ["I've got to rest before I fall apart. My joints are almost frozen. "],
- ["What a desolate place this is."],
- ["Where are you going?"],
- ["Well, I'm not going that way. It's much too rocky. This way is much easier."],
- ["What makes you think there are settlements over there?"],
- ["Don't get technical with me."],
- ["What mission? What are you talking about? I've had just about enough of you! Go that way! You'll be malfunctioning within a day, you nearsighted scrap pile!"],
- ["And don't let me catch you following me begging for help, because you won't get it."],
- ["No more adventures. I'm not going that way."],
- ["That malfunctioning little twerp. This is all his fault! He tricked me into going this way, but he'll do no better."],
- ["Wait, what's that? A transport! I'm saved!"],
- ["Over here! Help! Please, help!"],
- ["... so I cut off my power, shut down the afterburners and came in low on Deak's trail. I was so close I thought I was going to fry my instruments. As it was I busted up the Skyhopper pretty bad. Uncle Owen was pretty upset. He grounded me for the rest of the season. You should have been there... it was fantastic."],
- ["You ought to take it a little easy Luke. You may be the hottest bushpilot this side of Mos Eisley, but those little Skyhoppers are dangerous. Keep it up, and one day, whammo, you're going to be nothing more than a dark spot on the down side of a canyon wall."],
- ["Look who's talking. Now that you've been around those giant starships you're beginning to sound like my uncle. You've gotten soft in the city..."],
- ["I've missed you kid."],
- ["Well, things haven't been the same since you left, Biggs. It's been so...quiet."],
- ["Luke, I didn't come back just to say goodbye... I shouldn't tell you this, but you're the only one I can trust... and if I don't come back, I want somebody to know."],
- ["What are you talking about?"],
- ["I made some friends at the Academy. ... when our frigate goes to one of the central systems, we're going to jump ship and join the Alliance..."],
- ["Join the Rebellion?! Are you kidding! How?"],
- ["Quiet down will ya! You got a mouth bigger than a meteor crater!"],
- ["I'm sorry. I'm quiet. Listen how quiet I am. You can barely hear me..."],
- ["My friend has a friend on Bestine who might help us make contact."],
- ["around forever trying to find them."],
- ["I know it's a long shot, but if I don't find them I'll do what I can on my own... It's what we always talked about. Luke, I'm not going to wait for the Empire to draft me into service. The Rebellion is spreading and I want to be on the right side - the side I believe in. "],
- ["And I'm stuck here..."],
- ["I thought you were going to the Academy next term. You'll get your chance to get off this rock."],
- ["Not likely! I had to cancel my application. There has been a lot of unrest among the Sand People since you left... they've even raided the outskirts of Anchorhead."],
- ["Your uncle could hold off a whole colony of Sand People with one blaster."],
- ["I know, but he's got enough vaporators going to make the place pay off. He needs me for just one more season. I can't leave him now."],
- ["I feel for you, Luke, you're going to have to learn what seems to be important or what really is important. What good is all your uncle's work if it's taken over by the Empire?... You know they're starting to nationalize commerce in the central systems...it won't be long before your uncle is merely a tenant, slaving for the greater glory of the Empire."],
- ["It couldn't happen here. You said it yourself. The Empire won't bother with this rock."],
- ["Things always change."],
- ["I wish I was going... Are you going to be around long? "],
- ["No, I'm leaving in the morning..."],
- ["Then I guess I won't see you."],
- ["Maybe someday... I'll keep a lookout."],
- ["Well, I'll be at the Academy next season... after that who knows. I won't be drafted into the Imperial Starfleet that's for sure... Take care of yourself, you'll always be the best friend I've got."],
- ["So long, Luke."],
- ["Artoo-Detoo! It's you! It's you!"],
- ["Someone was in the pod. The tracks go off in this direction. "],
- ["Look, sir - droids."],
- ["Wake up! Wake up!"],
- ["We're doomed."],
- ["Do you think they'll melt us down?"],
- ["Don't shoot! Don't shoot! Will this never end?"],
- ["Luke, tell Owen that if he gets a translator to be sure it speaks Bocce."],
- ["It looks like we don't have much of a choice but I'll remind him."],
- ["I have no need for a protocol droid."],
- ["Sir - not in an environment such as this - that's why I've also been programmed for over thirty secondary functions that..."],
- ["What I really need is a droid that understands the binary languages of moisture vaporators."],
- ["Vaporators! Sir - My first job was programming binary load lifters... very similar to your vaporators. You could say..."],
- ["Do you speak Bocce?"],
- ["Of course I can, sir. It's like a second language for me... I'm as fluent in Bocce..."],
- ["All right; shut up! I'll take this one."],
- ["Shutting up, sir."],
- ["Luke, take these two over to the garage, will you? I want you to have both of them cleaned up before dinner."],
- ["But I was going into Toshi Station to pick up some power converters..."],
- ["You can waste time with your friends when your chores are done. Now, come on, get to it!"],
- ["All right, come on! And the red one, come on. Well, come on, Red, let's go."],
- ["Uncle Owen..."],
- ["Yeah?"],
- ["This R2 unit has a bad motivator. Look!"],
- ["Hey, what're you trying to push on us?"],
- ["Excuse me, sir, but that R2 unit is in prime condition. A real bargain."],
- ["Uncle Owen..."],
- ["Yeah?"],
- ["What about that one?"],
- ["What about that blue one? We'll take that one."],
- ["Yeah, take it away."],
- ["Uh, I'm quite sure you'll be very pleased with that one, sir. He really is in first-class condition. I've worked with him before. Here he comes."],
- ["Okay, let's go."],
- ["Now, don't forget this! Why I should stick my neck out for you is quite beyond my capacity!"],
- ["Thank the maker! This oil bath is going to feel so good. I've got such a bad case of dust contamination, I can barely move!"],
- ["It just isn't fair. Oh, Biggs is right. I'm never gonna get out of here!"],
- ["Is there anything I might do to help? "],
- ["Well, not unless you can alter time, speed up the harvest, or teleport me off this rock!"],
- ["I don't think so, sir. I'm only a droid and not very knowledgeable about such things. Not on this planet, anyways. As a matter of fact, I'm not even sure which planet I'm on."],
- ["Well, if there's a bright center to the universe, you're on the planet that it's farthest from."],
- ["I see, sir."],
- ["Uh, you can call me Luke."],
- ["I see, sir Luke."],
- ["Just Luke."],
- ["And I am See-Threepio, human-cyborg relations, and this is my counterpart, Artoo-Detoo."],
- ["Hello."],
- ["You got a lot of carbon scoring here. It looks like you boys have seen a lot of action."],
- ["With all we've been through, sometimes I'm amazed we're in as good condition as we are, what with the Rebellion and all."],
- ["You know of the Rebellion against the Empire?"],
- ["That's how we came to be in your service, if you take my meaning, sir."],
- ["Have you been in many battles?"],
- ["Several, I think. Actually, there's not much to tell. I'm not much more than an interpreter, and not very good at telling stories. Well, not at making them interesting, anyways."],
- ["Well, my little friend, you've got something jammed in here real good. Were you on a starcruiser or..."],
- ["Help me, Obi-Wan Kenobi. You'remy only hope."],
- ["What's this?"],
- ["What is what?!? He asked you a question...What is that?"],
- ["Help me, Obi-Wan Kenobi. You're my only hope. Help me, Obi-Wan Kenobi. You're my only hope."],
- ["Oh, he says it's nothing, sir. Merely a malfunction. Old data. Pay it no mind."],
- ["Who is she? She's beautiful."],
- ["I'm afraid I'm not quite sure, sir."],
- ["Help me, Obi-Wan Kenobi..."],
- ["I think she was a passenger on our last voyage. A person of some importance, sir - I believe. Our captain was attached to..."],
- ["Is there more to this recording?"],
- ["Behave yourself, Artoo. You're going to get us in trouble. It's all right, you can trust him. He's our new master."],
- ["He says he's the property of Obi-Wan Kenobi, a resident of these parts. And it's a private message for him. Quite frankly, sir, I don't know what he's talking about. Our last master was Captain Antilles, but with what we've been through, this little R2 unit has become a bit eccentric."],
- ["Obi-Wan Kenobi? I wonder if he means old Ben Kenobi?"],
- ["I beg your pardon, sir, but do you know what he's talking about?"],
- ["Well, I don't know anyone named Obi-Wan, but old Ben lives out beyond the dune sea. He's kind of a strange old hermit."],
- ["I wonder who she is. It sounds like she's in trouble. I'd better play back the whole thing."],
- ["He says the restraining bolt has short circuited his recording system. He suggests that if you remove the bolt, he might be able to play back the entire recording."],
- ["H'm? Oh, yeah, well, I guess you're too small to run away on me if I take this off! Okay."],
- ["There you go."],
- ["Well, wait a minute. Where'd she go? Bring her back! Play back the entire message."],
- ["been playing. The one you're carrying inside your rusty innards! "],
- ["Luke? Luke! Come to dinner!"],
- ["All right, I'll be right there, Aunt Beru."],
- ["I'm sorry, sir, but he appears to have picked up a slight flutter."],
- ["Well, see what you can do with him. I'll be right back."],
- ["Just you reconsider playing that message for him."],
- ["No, I don't think he likes you at all."],
- ["No, I don't like you either."],
- ["You know, I think that R2 unit we bought might have been stolen."],
- ["What makes you think that?"],
- ["Well, I stumbled across a recording while I was cleaning him. He says he belongs to someone called Obi-Wan Kenobi."],
- ["I thought he might have meant Ben. Do you know what he's talking about? Well, I wonder if he's related to Ben."],
- ["That old man's just a crazy wizard. Tomorrow I want you to take that R2 unit into Anchorhead and have its memory flushed. That'll be the end of it. It belongs to us now."],
- ["But what if this Obi-Wan comes looking for him?"],
- ["He won't, I don't think he exists any more. He died about the same time as your father."],
- ["He knew my father?"],
- ["I told you to forget it. Your only concern is to prepare the new droids for tomorrow. In the morning I want them on the south ridge working out those condensers."],
- ["Yes, sir. I think those new droids are going to work out fine. In fact, I, uh, was also thinking about our agreement about my staying on another season. And if these new droids do work out, I want to transmit my application to the Academy this year."],
- ["You mean the next semester before harvest?"],
- ["Sure, there're more than enough droids."],
- ["Harvest is when I need you the most. Only one more season. This year we'll make enough on the harvest so I'll be able to hire some more hands. And then you can go to the Academy next year."],
- ["You must understand I need you here, Luke."],
- ["But it's a whole 'nother year."],
- ["Look, it's only one more season."],
- ["Yeah, that's what you said last year when Biggs and Tank left."],
- ["Where are you going?"],
- ["It looks like I'm going nowhere. I have to finish cleaning those droids."],
- ["Owen, he can't stay here forever. Most of his friends have gone. It means so much to him."],
- ["I'll make it up to him next year. I promise."],
- ["Luke's just not a farmer, Owen. He has too much of his father in him."],
- ["That's what I'm afraid of."],
- ["What are you doing hiding there?"],
- ["It wasn't my fault, sir. Please don't deactivate me. I told him not to go, but he's faulty, malfunctioning; kept babbling on about his mission."],
- ["Oh, no!"],
- ["That R2 unit has always been a problem. These astro-droids are getting quite out of hand. Even I can't understand their logic at times. "],
- ["How could I be so stupid? He's nowhere in sight. Blast it!"],
- ["Pardon me, sir, but couldn't we go after him?"],
- ["It's too dangerous with all the Sand People around. We'll have to wait until morning."],
- ["Luke, I'm shutting the power down for the night."],
- ["All right, I'll be there in a few minutes. Boy, am I gonna get it."],
- ["You know that little droid is going to cause me a lot of trouble."],
- ["Oh, he excels at that, sir."],
- ["Luke? Luke? Luke? Where could he be loafing now!"],
- ["Have you seen Luke this morning?"],
- ["He said he had some things to do before he started today, so he left early."],
- ["Uh? Did he take those two new droids with him?"],
- ["I think so."],
- ["Well, he'd better have those units in the south range repaired bemidday or there'll be hell to pay!"],
- ["How's that."],
- ["Old Ben Kenobi lives out in this direction somewhere, but I don't see how that R2 unit could have come this far. We must have missed him. Uncle Owen isn't going to take this very well."],
- ["Sir, would it help if you told him it was my fault."],
- ["Sure. He needs you. He'd probably only deactivate you for a day or so..."],
- ["Deactivate! Well, on the other hand if you hadn't removed his restraining bolt..."],
- ["Wait, there's something dead ahead on the scanner. It looks like our droid... hit the accelerator."],
- ["Hey, whoa, just where do you think you're going?"],
- ["Master Luke here is your rightful owner. We'll have no more of this Obi-Wan Kenobi jibberish... and don't talk to me of your mission, either. You're fortunate he doesn't blast you into a million pieces right here."],
- ["Well, come on. It's getting late. I only hope we can get back before Uncle Owen really blows up."],
- ["If you don't mind my saying so, sir, I think you should deactivate the little fugitive until you've gotten him back to your workshop."],
- ["No, he's not going to try anything."],
- ["What's wrong with him now?"],
- ["Oh my... sir, he says there are several creatures approaching from the southeast."],
- ["Sand People! Or worse! Come on, let's have a look. Come on."],
- ["There are two Banthas down there but I don't see any... wait a second, they're Sand People all right. I can see one of them now."],
- ["Hello there! Come here my little friend. Don't be afraid."],
- ["Don't worry, he'll be all right."],
- ["What happened?"],
- ["Rest easy, son, you've had a busy day. You're fortunate you're still in one piece."],
- ["Ben? Ben Kenobi! Boy, am I glad to see you! "],
- ["The Jundland Wastes are not to be traveled lightly. Tell me, young Luke, what brings you out this far?"],
- ["Oh, this little droid! I think he's searching for his former master... I've never seen such devotion in a droid before... there seems tobe no stopping him. He claims to be the property of an Obi-Wan Kenobi. Is he a relative of yours? Do you know who he's talking about?"],
- ["Obi-Wan Kenobi... Obi-Wan? Now thats a name I haven't heard in a long time... a long time."],
- ["I think my uncle knew him. He said he was dead."],
- ["Oh, he's not dead, no... not yet."],
- ["You know him!"],
- ["Well of course, of course I know him. He's me! I haven't gone by the name Obi-Wan since oh, before you were born."],
- ["Then the droid does belong to you."],
- ["Don't seem to remember ever owning a droid. Very interesting... "],
- ["I think we better get indoors. The Sand People are easily startled but they will soon be back and in greater numbers."],
- ["... Threepio!"],
- ["Where am I? I must have taken a bad step..."],
- ["Can you stand? We've got to get out of here before the Sand People return."],
- ["I don't think I can make it. You go on, Master Luke. There's no sense in you risking yourself on my account. I'm done for."],
- ["No, you're not. What kind of talk is that?"],
- ["Quickly, son... they're on the move."],
- ["No, my father didn't fight in the wars. He was a navigator on a spice freighter."],
- ["That's what your uncle told you. He didn't hold with your father's ideals. Thought he should have stayed here and not gotten involved."],
- ["You fought in the Clone Wars?"],
- ["Yes, I was once a Jedi Knight the same as your father."],
- ["I wish I'd known him."],
- ["He was the best star-pilot in the galaxy, and a cunning warrior. I understand you've become quite a good pilot yourself. And he was a good friend. Which reminds me..."],
- ["I have something here for you. Your father wanted you to have this when you were old enough, but your uncle wouldn't allow it. He feared you might follow old Obi-Wan on some damned-fool idealistic crusade like your father did. "],
- ["Sir, if you'll not be needing me, I'll close down for awhile."],
- ["Sure, go ahead."],
- ["What is it?"],
- ["Your fathers lightsaber. This is the weapon of a Jedi Knight. Not as clumsy or as random as a blaster."],
- ["An elegant weapon for a morecivilized time. For over a thousand generations the Jedi Knights were the guardians of peace and justice in the Old Republic. Before the dark times, before the Empire."],
- ["How did my father die?"],
- ["A young Jedi named Darth Vader, who was a pupil of mine until he turned to evil, helped the Empire hunt down and destroy the Jedi Knights. He betrayed and murdered your father. Now the Jedi are all but extinct. Vader was seduced by the dark side of the Force."],
- ["The Force?"],
- ["Well, the Force is what gives the Jedi his power. It's an energy field created by all living things. It surrounds us and penetrates us. It binds the galaxy together."],
- ["Now, let's see if we can't figure out what you are, my little friend. And where you come from."],
- ["I saw part of the message he was..."],
- ["I seem to have found it."],
- ["General Kenobi, years ago you served my father in the Clone Wars. Now he begs you to help him in his struggle against the Empire. I regret that I am unable to present my father's request to you in person, but my ship has fallen under attack and I'm afraid my mission to bring you to Alderaan has failed. I have placed information vital to the survival of the Rebellion into the memory systems of this R2 unit. My father will know how to retrieve it. You must see this droid safely delivered to him on Alderaan. This is our most desperate hour. Help me, Obi-Wan Kenobi, you're my only hope."],
- ["You must learn the ways of the Force if you're to come with me to Alderaan."],
- ["Alderaan? I'm not going to Alderaan. I've got to go home. It's late, I'm in for it as it is."],
- ["I need your help, Luke. I'm getting too old for this sort of thing.She needs your help."],
- ["I can't get involved! I've got work to do! It's not that I like the Empire. I hate it! But there's nothing I can do about it right now. It's such a long way from here."],
- ["That's your uncle talking."],
- ["Oh, God, my uncle. How am I ever going to explain this?"],
- ["Learn about the Force, Luke."],
- ["Look, I can take you as far as Anchorhead. You can get a transport there to Mos Eisley or wherever you're going."],
- ["You must do what you feel is right, of course."],
- ["Until this battle station is fully operational we are vulnerable. The Rebel Alliance is too well equipped. They're more dangerous than you realize."],
- ["Dangerous to your starfleet, Commander; not to this battle station!"],
- ["The Rebellion will continue to gain support in the Imperial Senate as long as...."],
- ["The Imperial Senate will no longer be of any concern to us. I've just received word that the Emperor has dissolved the council permanently. The last remnants of the Old Republic have been swept away."],
- ["That's impossible! How will the Emperor maintain control without the bureaucracy?"],
- ["The regional governors now have direct control over territories. Fear will keep the local systems in line. Fear of this battle station."],
- ["And what of the Rebellion? If the Rebels have obtained a complete technical readout of this station, it is possible, however unlikely, that they might find a weakness and exploit it."],
- ["The plans you refer to will soon be back in our hands."],
- ["Any attack made by the Rebels against this station would be a useless gesture, no matter what technical data they've obtained. This station is now the ultimate power in the universe. I suggest we use it!"],
- ["Don't be too proud of this technological terror you've constructed. The ability to destroy a planet is insignificant next to the power of the Force."],
- ["Don't try to frighten us with your sorcerer's ways, Lord Vader. Your sad devotion to that ancient religion has not helped you conjure up the stolen data tapes, or given you clairvoyance enough to find the Rebel's hidden fort..."],
- ["I find your lack of faith disturbing."],
- ["Enough of this! Vader, release him!"],
- ["As you wish."],
- ["This bickering is pointless. Lord Vader will provide us with the location of the Rebel fortress by the time this station is operational. We will then crush the Rebellion with one swift stroke."],
- ["It looks like Sand People did this, all right. Look, here are gaffi sticks, bantha tracks. It's just... I never heard of them hitting anything this big before."],
- ["They didn't. But we are meant to think they did. These tracks are side by side. Sand People always ride single file to hide there numbers."],
- ["These are the same Jawas that sold us Artoo and Threepio."],
- ["And these blast points, too accurate for Sand People. Only Imperial stormtroopers are so precise."],
- ["Why would Imperial troops want to slaughter Jawas?"],
- ["If they traced the robots here, they may have learned who they sold them to. And that would lead them back home!"],
- ["Wait, Luke! It's too dangerous."],
- ["Uncle Owen! Aunt Beru! Uncle Owen!"],
- ["And, now Your Highness, we will discuss the location of your hidden Rebel base."],
- ["There's nothing you could have done, Luke, had you been there. You'd have been killed, too, and the droids would now be in the hands of the Empire."],
- ["I want to come with you to Alderaan. There's nothing here for me now. I want to learn the ways of the Force and become a Jedi like my father."],
- ["Mos Eisley Spaceport. You will never find a more wretched hive of scum and villainy. We must be cautious."],
- ["How long have you had these droids?"],
- ["About three or four seasons."],
- ["They're for sale if you want them."],
- ["Let me see your identification."],
- ["You don't need to see his identification."],
- ["We don't need to see his identification."],
- ["looking for."],
- ["These are not the droids we're looking for."],
- ["He can go about his business."],
- ["You can go about your business."],
- ["Move along."],
- ["Move along. Move along."],
- ["I can't abide these Jawas. Disgusting creatures."],
- ["Go on, go on. I can't understand how we got by those troopers. I thought we were dead."],
- ["The Force can have a strong influence on the weak-minded. You will find it a powerful ally."],
- ["Do you really think we're going to find a pilot here that'll take us to Alderaan?"],
- ["Well, most of the best freighter pilots can be found here. Only watch your step. This place can be a little rough."],
- ["I'm ready for anything."],
- ["Come along, Artoo."],
- ["We don't serve their kind here!"],
- ["What?"],
- ["Your droids. They'll have to wait outside. We don't want them here."],
- ["Listen, why don't you wait out by the speeder. We don't want any trouble."],
- ["I heartily agree with you sir."],
- ["Negola dewaghi wooldugger?!?"],
- ["He doesn't like you."],
- ["I'm sorry."],
- ["I don't like you either"],
- ["Don't insult us. You just watch yourself. We're wanted men. I have the death sentence on twelve systems."],
- ["I'll be careful than."],
- ["You'll be dead."],
- ["This little one isn't worth the effort. Come let me buy you something..."],
- ["No blasters! No blaster!"],
- ["This is Chewbacca. He's first-mate on a ship that might suit our needs."],
- ["I don't like the look of this."],
- ["Han Solo. I'm captain of the Millennium Falcon. Chewie here tells me you're looking for passage to the Alderaan system."],
- ["Yes, indeed. If it's a fast ship."],
- ["Fast ship? You've never heard of the Millennium Falcon?"],
- ["Should I have?"],
- ["It's the ship that made the Kessel run in less than twelve parsecs!"],
- ["I've outrun Imperial starships, not the local bulk-cruisers, mind you. I'm talking about the big Corellian ships now. She's fast enough for you, old man. What's the cargo?"],
- ["Only passengers. Myself, the boy, two droids, and no questions asked."],
- ["What is it? Some kind of local trouble?"],
- ["Let's just say we'd like to avoid any Imperial entanglements."],
- ["Well, that's the real trick, isn't it? And it's going to cost you something extra. Ten thousand in advance."],
- ["Ten thousand? We could almost buy our own ship for that!"],
- ["But who's going to fly it, kid! You?"],
- ["You bet I could. I'm not such a bad pilot myself! We don't have to sit here and listen..."],
- ["We haven't that much with us. But we could pay you two thousand now, plus fifteen when we reach Alderaan."],
- ["Seventeen, huh!"],
- ["Okay. You guys got yourself a ship. We'll leave as soon as you're ready. Docking bay Ninety-four."],
- ["Ninety-four."],
- ["Looks like somebody's beginning to take an interest in your handiwork."],
- ["All right, we'll check it out."],
- ["Seventeen thousand! Those guys must really be desperate. This could really save my neck. Get back to the ship and get her ready."],
- ["You'll have to sell your speeder."],
- ["That's okay. I'm never coming back to this planet again."],
- ["Going somewhere, Solo?"],
- ["Yes, Greedo. As a matter of fact, I was just going to see your boss. Tell Jabba that I've got his money."],
- ["It's too late. You should have paid him when you had the chance. Jabba's put a price on your head, so large that every bounty hunter in the galaxy will be looking for you. I'm lucky I found you first."],
- ["Yeah, but this time I got the money."],
- ["If you give it to me, I might forget I found you."],
- ["I don't have it with me. Tell Jabba..."],
- ["Jabba's through with you. He has no time for smugglers who drop their shipments at the first sign of an Imperial cruiser."],
- ["Even I get boarded sometimes. Do you think I had a choice?"],
- ["You can tell that to Jabba. He may only take your ship."],
- ["Over my dead body."],
- ["That's the idea I've been looking forward to killing you for a long time."],
- ["Yes, I'll bet you have."],
- ["Sorry about the mess."],
- ["Her resistance to the mind probe is considerable. It will be some time before we can extract any information from her."],
- ["The final check-out is complete. All systems are operational. What course shall we set?"],
- ["Perhaps she would respond to an alternative form of persuasion."],
- ["What do you mean?"],
- ["I think it is time we demonstrate the full power of this station.Set your course for Princess Leia's home planet of Alderaan."],
- ["With pleasure."],
- ["Lock the door, Artoo."],
- ["All right, check that side of the street. It's secure. Move on to the next door."],
- ["I would much rather have gone with Master Luke than stay here with you. I don't know what all this trouble is about, but I'm sure it must be your fault."],
- ["You watch your language!"],
- ["He says it's the best he can do. Since the XP-38 came out, they "],
- ["It will be enough."],
- ["If the ship's as fast as he's boasting, we ought to do well."],
- ["Come on out, Solo!"],
- ["I've been waiting for you, Jabba."],
- ["I expected you would be."],
- ["I'm not the type to run."],
- ["Han, my boy, there are times when you disappoint me... why haven't you paid me? And why did you have to fry poor Greedo like that... after all we've been through together."],
- ["You sent Greedo to blast me."],
- ["Han, why you're the best smuggler in the business. You're too valuable to fry. He was only relaying my concern at your delays. He wasn't going to blast you."],
- ["I think he thought he was. Next time don't send one of those twerps. If you've got something to say to me, come see me yourself."],
- ["Han, Han! If only you hadn't had to dump that shipment of spice... you understand I just can't make an exception. Where would I be if every pilot who smuggled for me dumped their shipment at the first sign of an Imperial starship? It's not good business."],
- ["You know, even I get boarded sometimes, Jabba. I had no choice, but I've gota charter now and I can pay you back, plus a little extra. I just need some more time."],
- ["Put your blasters away. Han, my boy, I'm only doing this because you're the best and I need you. So, for an extra, say... twenty percent I'll give you a little more time... but this is it. If you disappoint me again, I'll put a price on your head so large you won't be able to go near a civilized system for the rest of your short life."],
- ["Jabba, I'll pay you because it's my pleasure."],
- ["What a piece of junk."],
- ["She'll make point five beyond the speed of light. She may not look like much, but she's got it where it counts, kid. I've added some special modifications myself."],
- ["We're a little rushed, so if you'll hurry aboard we'll get out of here."],
- ["Hello, sir."],
- ["Which way?"],
- ["All right, men. Load your weapons!"],
- ["Stop that ship!"],
- ["Blast 'em!"],
- ["Chewie, get us out of here!"],
- ["Oh, my. I'd forgotten how much I hate space travel."],
- ["It looks like an Imperial cruiser. Our passengers must be hotter than I thought. Try and hold them off. Angle the deflector shield while I make the calculations for the jump to light speed."],
- ["Stay sharp! There are two more coming in; they're going to try to cut us off."],
- ["Why don't you outrun them? I thought you said this thing was fast."],
- ["Watch your mouth, kid, or you're going to find yourself floating home. We'll be safe enough once we make the jump to hyperspace. Besides, I know a few maneuvers. We'll lose them!"],
- ["Here's where the fun begins!"],
- ["How long before you can make the jump to light speed?"],
- ["It'll take a few moments to get the coordinates from the navi-computer."],
- ["Are you kidding? At the rate they're gaining..."],
- ["Traveling through hyperspace isn't like dusting crops, boy! Without precise calculations we could fly right through a star or bounce too close to a supernova and that'd end your trip real quick, wouldn't it?"],
- ["What's that flashing?"],
- ["We're losing our deflector shield. Go strap yourself in, I'm going to make the jump to light speed."],
- ["We've entered the Alderaan system."],
- ["Governor Tarkin, I should have expected to find you holding Vader's leash. I recognized your foul stench when I was brought on board."],
- ["Charming to the last. You don't know how hard I found it signing the order to terminate your life!"],
- ["to take the responsibility yourself!"],
- ["Princess Leia, before your execution I would like you to be my guest at a ceremony that will make this battle station operational. No star system will dare oppose the Emperor now."],
- ["The more you tighten your grip, Tarkin, the more star systems will slip through your fingers."],
- ["Not after we demonstrate the power of this station. In a way, you have determined the choice of the planet that'll be destroyed first. Since you are reluctant to provide us with the location of the Rebel base, I have chosen to test this station's destructive power... on your home planet of Alderaan."],
- ["No! Alderaan is peaceful. We have no weapons. You can't possibly..."],
- ["You would prefer another target? A military target? Then name the system!"],
- ["I grow tired of asking this. So it'll be the last time. Where is the Rebel base?"],
- ["Dantooine."],
- ["They're on Dantooine."],
- ["There. You see Lord Vader, she can be reasonable. Continue with the operation. You may fire when ready."],
- ["What?"],
- ["You're far too trusting. Dantooine is too remote to make an effective demonstration. But don't worry. We will deal with your Rebel friends soon enough. "],
- ["No!"],
- ["Commence primary ignition."],
- ["Are you all right? What's wrong?"],
- ["I felt a great disturbance in the Force... as if millions of voices suddenly cried out in terror and were suddenly silenced. I fear something terrible has happened."],
- ["You'd better get on with your exercises."],
- ["Well, you can forget your troubles with those Imperial slugs. I told you I'd outrun 'em."],
- ["Don't everyone thank me at once."],
- ["Anyway, we should be at Alderaan about oh-two-hundred hours."],
- ["Now be careful, Artoo."],
- ["He made a fair move. Screaming about it won't help you."],
- ["Let him have it. It's not wise to upset a Wookiee."],
- ["But sir, nobody worries about upsetting a droid."],
- ["That's 'cause droids don't pull people's arms out of their socket when they lose. Wookiees are known to do that."],
- ["I see your point, sir. I suggest a new strategy, Artoo. Let the Wookiee win."],
- ["Remember, a Jedi can feel the Force flowing through him."],
- ["You mean it controls your actions?"],
- ["Partially. But it also obeys your commands."],
- ["Hokey religions and ancient weapons are no match for a good blaster at your side, kid."],
- ["You don't believe in the Force, do you?"],
- ["Kid, I've flown from one side of this galaxy to the other. I've seen a lot of strange stuff, but I've never seen anything to make me believe there's one all-powerful force controlling everything. There's no mystical energy field that controls my destiny."],
- ["It's all a lot of simple tricks and nonsense."],
- ["I suggest you try it again, Luke."],
- ["This time, let go your conscious self and act on instinct."],
- ["With the blast shield down, I can't even see. How am I supposed to fight?"],
- ["Your eyes can deceive you. Don't trust them."],
- ["Stretch out with your feelings."],
- ["You see, you can do it."],
- ["I call it luck. "],
- ["In my experience, there's no such thing as luck."],
- ["Look, going good against remotes is one thing. Going good against the living? That's something else."],
- ["Looks like we're coming up on Alderaan."],
- ["You know, I did feel something. I could almost see the remote."],
- ["That's good. You have taken your first step into a larger world."],
- ["Yes."],
- ["Our scout ships have reached Dantooine. They found the remains of a Rebel base, but they estimate that it has been deserted for some time. They are now conducting an extensive search of the surrounding systems."],
- ["She lied! She lied to us!"],
- ["I told you she would never consciously betray the Rebellion."],
- ["Terminate her... immediately!"],
- ["Stand by, Chewie, here we go. Cut in the sublight engines."],
- ["What the...? Aw, we've come out of hyperspace into a meteor shower. Some kind of asteroid collision. It's not on any of the charts."],
- ["What's going on?"],
- ["Our position is correct, except... no, Alderaan!"],
- ["What do you mean? Where is it?"],
- ["Thats what I'm trying to tell you, kid. It ain't there. It's been totally blown away."],
- ["What? How?"],
- ["Destroyed... by the Empire!"],
- ["The entire starfleet couldn't destroy the whole planet. It'd take a thousand ships with more fire power than I've..."],
- ["There's another ship coming in."],
- ["Maybe they know what happened."],
- ["It's an Imperial fighter."],
- ["It followed us!"],
- ["No. It's a short range fighter."],
- ["There aren't any bases around here. Where did it come from?"],
- ["It sure is leaving in a big hurry. If they identify us, we're in big trouble."],
- ["Not if I can help it. Chewie... jam it's transmissions."],
- ["It'd be as well to let it go. It's too far out of range."],
- ["Not for long..."],
- ["A fighter that size couldn't get this deep into space on its own."],
- ["Then he must have gotten lost, been part of a convoy or something..."],
- ["Well, he ain't going to be around long enough to tell anyone about us."],
- ["Look at him. He's heading for that small moon."],
- ["I think I can get him before he gets there... he's almost in range."],
- ["That's no moon! It's a space station."],
- ["It's too big to be a space station."],
- ["I have a very bad feeling about this."],
- ["Turn the ship around!"],
- ["Yeah, I think your right. Full reverse! Chewie, lock in the auxiliary power."],
- ["Why are we still moving towards it?"],
- ["We're caught in a tractor beam! It's pulling us in!"],
- ["But there's gotta be something you can do!"],
- ["There's nothin' I can do about it, kid. I'm in full power. I'm going to have to shut down. But they're not going to get me without a fight!"],
- ["You can't win. But there are alternatives to fighting."],
- ["Clear Bay twenty-three-seven. We are opening the magnetic field."],
- ["To your stations!"],
- ["Come with me."],
- ["Close all outboard shields! Close all outboard shields!"],
- ["Yes."],
- ["We've captured a freighter entering the remains of the Alderaan system. It's markings match those of a ship that blasted its way out of Mos Eisley."],
- ["They must be trying to return the stolen plans to the princess. She may yet be of some use to us."],
- ["Unlock one-five-seven and nine. Release charges."],
- ["There's no one on board, sir. According to the log, the crew abandoned ship right after takeoff. It must be a decoy, sir. Several of the escape pods have been jettisoned."],
- ["Did you find any droids?"],
- ["No, sir. If there were any on board, they must also have jettisoned."],
- ["Send a scanning crew on board. I want every part of this ship checked."],
- ["Yes, sir."],
- ["I sense something... a presence I haven't felt since..."],
- ["Get me a scanning crew in here on the double. I want every part of this ship checked!"],
- ["Boy, it's lucky you had these compartments."],
- ["I use them for smuggling. I never thought I'd be smuggling myself in them. This is ridiculous. Even if I could take off, I'd never get past the tractor beam."],
- ["Leave that to me!"],
- ["Damn fool. I knew that you were going to say that!"],
- ["Who's the more foolish... the fool or the fool who follows him?"],
- ["The ship's all yours. If the scanners pick up anything, report it immediately. All right, let's go."],
- ["Hey down there, could you give us a hand with this?"],
- ["TX-four-one-two. Why aren't you at your post? TX-four-one-two, do you copy? "],
- ["Take over. We've got a bad transmitter. I'll see what I can do."],
- ["You know, between his howling and your blasting everything in sight, it's a wonder the whole station doesn't know we're here."],
- ["Bring them on! I prefer a straight fight to all this sneaking around."],
- ["We found the computer outlet, sir."],
- ["Plug in. He should be able to interpret the entire Imperial computer network."],
- ["He says he's found the main control to the power beam that's holding the ship here. He'll try to make the precise location appear on the monitor."],
- ["The tractor beam is coupled to the main reactor in seven locations. A power loss at one of the terminals will allow the ship to leave."],
- ["I don't think you boys can help. I must go alone."],
- ["Whatever you say. I've done more than I bargained for on this trip already."],
- ["I want to go with you."],
- ["Be patient, Luke. Stay and watch over the droids."],
- ["But he can..."],
- ["They must be delivered safely or other star systems will suffer the same fate as Alderaan. Your destiny lies along a different path from mine. The Force will be with you... always!"],
- ["Boy you said it, Chewie."],
- ["Where did you dig up that old fossil?"],
- ["Ben is a great man."],
- ["Yeah, great at getting us into trouble."],
- ["I didn't hear you give any ideas..."],
- ["Well, anything would be better than just hanging around waiting for them to pick us up..."],
- ["Who do you think..."],
- ["What is it?"],
- ["I'm afraid I'm not quite sure, sir. He says "],
- ["Well, who... who has he found?"],
- ["Princess Leia."],
- ["The princess? She's here?"],
- ["Princess?"],
- ["Where... where is she?"],
- ["Princess? What's going on?"],
- ["Level five. Detention block AA-twenty-three. I'm afraid she's scheduled to be terminated."],
- ["Oh, no! We've got to do something."],
- ["What are you talking about?"],
- ["The droid belongs to her. She's the one in the message. We've got to help her."],
- ["Now, look, don't get any funny ideas. The old man wants us to wait right here."],
- ["But he didn't know she was here. Look, will you just find a way back into the detention block?"],
- ["I'm not going anywhere."],
- ["They're going to execute her. Look, a few minutes ago you said you didn't want to just wait here to be captured. Now all you want to do is stay. "],
- ["Marching into the detention area is not what I had in mind."],
- ["But they're going to kill her!"],
- ["Better her than me..."],
- ["She's rich."],
- ["Rich?"],
- ["Yes. Rich, powerful! Listen, if you were to rescue her, the reward would be..."],
- ["What?"],
- ["Well more wealth that you can imagine."],
- ["I don't know, I can imagine quite a bit!"],
- ["You'll get it!"],
- ["I better!"],
- ["You will..."],
- ["All right, kid. But you'd better be right about this!"],
- ["All right."],
- ["What's your plan?"],
- ["Uh... Threepio, hand me those binders there will you?"],
- ["Okay. Now, I'm going to put these on you."],
- ["Okay. Han, you put these on."],
- ["Don't worry, Chewie. I think I know what he has in mind."],
- ["Master Luke, sir! Pardon me for asking... but, ah... what should Artoo and I do if we're discovered here?"],
- ["Lock the door!"],
- ["And hope they don't have blasters."],
- ["That isn't very reassuring."],
- ["I can't see a thing in this helmet."],
- ["This is not going to work."],
- ["Why didn't you say so before?"],
- ["I did say so before!"],
- ["Where are you taking this... thing?"],
- ["Prisoner transfer from Block one-one-three-eight."],
- ["I wasn't notified. I'll have to clear it."],
- ["Look out! He's loose!"],
- ["He's going to pull us all apart."],
- ["Go get him!"],
- ["We've got to find out which cell this princess of yours is in. Here it is... cell twenty-one-eight-seven. You go get her. I'll hold them here."],
- ["Everything is under control. Situation normal."],
- ["What happened?"],
- ["Uh... had a slight weapons malfunction. But, uh, everything's perfectly all right now. We're fine. We're all fine here, now, thank you. How are you?"],
- ["We're sending a squad up."],
- ["Uh, uh, negative, negative. We had a reactor leak here now. Give us a few minutes to lock it down. Large leak... very dangerous."],
- ["Who is this? What's your operating number?"],
- ["Boring conversation anyway.Luke! We're going to have company!"],
- ["Aren't you a little short for a stormtrooper?"],
- ["What? Oh... the uniform. I'm Luke Skywalker. I'm here to rescue you. "],
- ["You're who?"],
- ["I'm here to rescue you. I've got your R2 unit. I'm here with Ben Kenobi."],
- ["Ben Kenobi is here! Where is he?"],
- ["Come on!"],
- ["He is here..."],
- ["Obi-Wan Kenobi! What makes you think so?"],
- ["A tremor in the Force. The last time I felt it was in the presence of my old master."],
- ["Surely he must be dead by now."],
- ["Don't underestimate the Force."],
- ["The Jedi are extinct, their fire has gone out of the universe. You, my friend, are all that's left of their religion."],
- ["Yes."],
- ["Governor Tarkin, we have an emergency alert in detention block AA-twenty-three."],
- ["The princess! Put all sections on alert!"],
- ["Obi-Wan is here. The Force is with him."],
- ["If you're right, he must not be allowed to escape."],
- ["Escape may not his plan. I must face him alone."],
- ["Chewie!"],
- ["Get behind me! Get behind me!"],
- ["Can't get out that way."],
- ["Looks like you managed to cut off our only escape route."],
- ["Maybe you'd like it back in your cell, Your Highness."],
- ["See-Threepio! See-Threepio!"],
- ["Yes sir?"],
- ["We've been cut off! Are there any other ways out of the cell bay?...What was that? I didn't copy!"],
- ["I said, all systems have been alerted to your presence, sir. The main entrance seems to be the only way out; all other information on your level is restricted."],
- ["Open up in there!"],
- ["Oh, no!"],
- ["There isn't any other way out."],
- ["I can't hold them off forever! Now what?"],
- ["This is some rescue. When you came in here, didn't you have a plan for getting out?"],
- ["He's the brains, sweetheart."],
- ["Well, I didn't..."],
- ["What the hell are you doing?"],
- ["Somebody has to save our skins. Into the garbage chute, wise guy."],
- ["Get in there you big furry oaf! I don't care what you smell! Get in there and don't worry about it."],
- ["Wonderful girl! Either I'm going to kill her or I'm beginning to like her. Get in there!"],
- ["Oh! The garbage chute was a really wonderful idea. What an incredible smell you've discovered! Let's get out of here! Get away from there..."],
- ["No! wait!"],
- ["Will you forget it? I already tried it. It's magnetically sealed!"],
- ["Put that thing away! You're going to get us all killed."],
- ["Absolutely, Your Worship. Look, I had everything under control until you led us down here. You know, it's not going to take them long to figure out what happened to us."],
- ["It could be worse..."],
- ["It's worse."],
- ["There's something alive in here!"],
- ["That's your imagination."],
- ["Something just moves past my leg! Look! Did you see that?"],
- ["What?"],
- ["Help!"],
- ["Luke! Luke! Luke!"],
- ["Luke!"],
- ["Luke, Luke, grab a hold of this."],
- ["Blast it, will you! My gun's jammed."],
- ["Where?"],
- ["Anywhere! Oh!!"],
- ["Luke! Luke!"],
- ["Grab him!"],
- ["What happened?"],
- ["I don't know, it just let go of me and disappeared..."],
- ["I've got a very bad feeling about this."],
- ["The walls are moving!"],
- ["Don't just stand there. Try and brace it with something."],
- ["Wait a minute!"],
- ["Threepio! Come in Threepio! Threepio! Where could he be?"],
- ["Take over!See to him! Look there!"],
- ["They're madmen! They're heading for the prison level. If you hurry, you might catch them."],
- ["Follow me! You stand guard."],
- ["Come on!"],
- ["Oh! All this excitement has overrun the circuits of my counterpart here. If you don't mind, I'd like to take him down to maintenance."],
- ["All right."],
- ["Threepio! Come in, Threepio! Threepio!"],
- ["Get to the top!"],
- ["I can't "],
- ["Where could he be? Threepio! Threepio, will you come in?"],
- ["They aren't here! Something must have happened to them. See if they've been captured."],
- ["Hurry!"],
- ["One thing's for sure. We're all going to be a lot thinner!Get on top of it!"],
- ["I'm trying!"],
- ["Thank goodness, they haven't found them! Where could they be?"],
- ["Use the comlink? Oh, my! I forgot I turned it off!"],
- ["Are you there, sir?"],
- ["Threepio!"],
- ["We've had some problems..."],
- ["Shut down all the garbage mashers on the detention level, will you? Do you copy?"],
- ["Shut down all the garbage mashers on the detention level."],
- ["Shut down all the garbage mashers on the detention level."],
- ["No. Shut them all down! Hurry!"],
- ["Listen to them! They're dying, Artoo! Curse my metal body! I wasn't fast enough. It's all my fault! My poor master!"],
- ["Threepio, we're all right!"],
- ["We're all right. You did great."],
- ["Hey... hey, open the pressure maintenance hatch on unit number... where are we?"],
- ["Three-two-six-eight-two-seven."],
- ["If we can just avoid any more female advice, we ought to be able to get out of here."],
- ["Well, let's get moving!"],
- ["Where are you going?"],
- ["No, wait. They'll hear!"],
- ["Come here, you big coward!"],
- ["Chewie! Come here!"],
- ["Listen. I don't know who you are, or where you came from, but from now on, you do as I tell you. Okay?"],
- ["Look, Your Worshipfulness, let's get one thing straight! I takeorders from one person! Me!"],
- ["It's a wonder you're still alive.Will somebody get this big walking carpet out of my way?"],
- ["No reward is worth this."],
- ["Secure this area until the alert is canceled."],
- ["Give me regular reports."],
- ["Do you know what's going on?"],
- ["Maybe it's another drill."],
- ["What was that?"],
- ["Oh, it's nothing. Don't worry about it."],
- ["There she is."],
- ["See-Threepio, do you copy?"],
- ["For the moment. Uh, we're in the main hangar across from the ship."],
- ["We're right above you. Stand by."],
- ["You came in that thing? You're braver that I thought."],
- ["Nice! Come on!"],
- ["It's them! Blast them!"],
- ["Get back to the ship!"],
- ["Where are you going? Come back!"],
- ["He certainly has courage."],
- ["What good will it do us if he gets himself killed? Come on!"],
- ["I think we took a wrong turn."],
- ["There's no lock!"],
- ["That oughta hold it for a while."],
- ["Quick, we've got to get across. Find the control that extends the bridge."],
- ["Oh, I think I just blasted it."],
- ["They're coming through!"],
- ["Here, hold this."],
- ["Here they come!"],
- ["For luck!"],
- ["Where could they be?"],
- ["Close the blast doors!"],
- ["Open the blast doors! Open the blast doors!"],
- ["I've been waiting for you, Obi-Wan. We meet again, at last. The circle is now complete."],
- ["When I left you, I was but the learner; now I am the master."],
- ["Only a master of evil, Darth."],
- ["Your powers are weak, old man."],
- ["You can't win, Darth. If you strike me down, I shall become more powerful than you can possibly imagine."],
- ["Didn't we just leave this party?"],
- ["What kept you?"],
- ["We ran into some old friends."],
- ["Is the ship all right?"],
- ["Seems okay, if we can get to it.Just hope the old man got the tractor beam out of commission."],
- ["Look!"],
- ["Come on, Artoo, we're going!"],
- ["Now's our chance! Go!"],
- ["No!"],
- ["Come on!"],
- ["Come on! Luke, its too late!"],
- ["Blast the door! Kid!"],
- ["Run, Luke! Run!"],
- ["I hope the old man got that tractor beam out if commission, or this is going to be a real short trip. Okay, hit it!"],
- ["We're coming up on the sentry ships. Hold 'em off! Angle the deflector shields while I charge up the main guns!"],
- ["I can't believe he's gone."],
- ["There wasn't anything you could have done."],
- ["Come on, buddy, we're not out of this yet!"],
- ["You in, kid? Okay, stay sharp!"],
- ["Here they come!"],
- ["They're coming in too fast!"],
- ["Oooh!"],
- ["We've lost lateral controls."],
- ["Don't worry, she'll hold together."],
- ["You hear me, baby? Hold together!"],
- ["Got him! I got him!"],
- ["Great kid! Don't get cocky."],
- ["There are still two more of them out there!"],
- ["That's it! We did it!"],
- ["We did it!"],
- ["Help! I think I'm melting!This is all your fault."],
- ["Are they away?"],
- ["They have just made the jump into hyperspace."],
- ["You're sure the homing beacon is secure aboard their ship? I'm taking an awful risk, Vader. This had better work."],
- ["Not a bad bit of rescuing, huh? You know, sometimes I even amaze myself."],
- ["That doesn't sound too hard. Besides, they let us go. It's the only explanation for the ease of our escape."],
- ["Easy... you call that easy?"],
- ["Their tracking us!"],
- ["Not this ship, sister."],
- ["At least the information in Artoo is still intact."],
- ["What's so important? What's he carrying?"],
- ["The technical readouts of that battle station. I only hope that when the data is analyzed, a weakness can be found. It's not over yet!"],
- ["It is for me, sister! Look, I ain't in this for your revolution, and I'm not in it for you, Princess. I expect to be well paid. I'm in it for the money!"],
- ["You needn't worry about your reward. If money is all that you love, then that's what you'll receive!"],
- ["Your friend is quite a mercenary. I wonder if he really cares about anything... or anyone."],
- ["I care!"],
- ["So... what do you think of her, Han?"],
- ["I'm trying not to, kid!"],
- ["Good..."],
- ["Still, she's got a lot of spirit. I don't know, what do you think? Do you think a princess and a guy like me..."],
- ["No!"],
- ["You're safe! We had feared the worst."],
- ["When we heard about Alderaan, we were afraid that you were... lost along with your father."],
- ["We don't have time for our sorrows, commander. The battle station has surely tracked us here.It's the only explanation for the ease of our escape. You must use the information in this R2 unit to plan the attack. It is our only hope."],
- ["Yes."],
- ["We are approaching the planet Yavin. The Rebel base is on a moon on the far side. We are preparing to orbit the planet."],
- ["The battle station is heavily shielded and carries a firepower greater than half the star fleet.Its defenses are designed around a direct large-scale assault. A small one-man fighter should be able to penetrate the outer defense."],
- ["Pardon me for asking, sir, but what good are snub fighters going to be against that?"],
- ["Well, the Empire doesn't consider a small one-man fighter to be any threat, or they'd have a tighter defense. An analysis of the plans provided by Princess Leia has demonstrated a weakness in the battle station."],
- ["The approach will not be easy. You are required to maneuver straight down this trench and skim the surface to this point. The target area is only two meters wide. It's a small thermal exhaust port, right below the main port. The shaft leads directly to the reactor system. A precise hit will start a chain reaction which should destroy the station."],
- ["Only a precise hit will set up a chain reaction. The shaft is ray-shielded, so you'll have to use proton torpedoes."],
- ["That's impossible, even for a computer."],
- ["It's not impossible. I used to bullseye womp rats in my T-sixteen back home. They're not much bigger than two meters."],
- ["Man your ships! And may the Force be with you!"],
- ["Orbiting the planet at maximum velocity. The moon with the Rebel base will be in range in thirty minutes."],
- ["This will be a day long remembered. It has seen the end of Kenobi and it will soon see the end of the Rebellion."],
- ["All flight troops, man your stations. All flight troops, man your stations."],
- ["So... you got your reward and you're just leaving then?"],
- ["That's right, yeah! I got some old debts I've got to pay off with this stuff. Even if I didn't, you don't think I'd be fool enough to stick around here, do you? Why don't you come with us? You're pretty good in a fight. I could use you."],
- ["Come on! Why don't you take a look around? You know what's about to happen, what they're up against. They could use a good pilot like you. You're turning your back on them."],
- ["What good's a reward if you ain't around to use it? Besides, attacking that battle station ain'tmy idea of courage. It's more like suicide."],
- ["All right. Well, take care of yourself, Han... guess that's what you're best at, isn't it?"],
- ["Hey, Luke... may the Force be with you!"],
- ["What're you lookin' at? I know what I'm doing."],
- ["What's wrong?"],
- ["Oh, it's Han! I don't know, I really thought he'd change his mind. "],
- ["He's got to follow his own path. No one can choose it for him."],
- ["I only wish Ben were here."],
- ["Luke! I don't believe it! How'd you get here... are you going out with us?!"],
- ["Biggs! Of course, I'll be up there with you! Listen, have I got some stories to tell..."],
- ["Are you... Luke Skywalker? Have you been checked out on the Incom T-sixty-five?"],
- ["Sir, Luke is the best bushpilot in the outer rim territories."],
- ["I met your father once when I was just a boy. He was a great pilot. You'll do all right. If you've got half of your father's skill, you'll do better than all right."],
- ["Thank you, sir. I'll try."],
- ["I've got to get aboard. Listen, you'll tell me your stories when we come back. All right?"],
- ["I told you I'd make it someday, Biggs."],
- ["You did, all right. It's going to be like old times Luke. We're a couple of shooting stars that'll never be stopped!"],
- ["This R2 unit of your seems a bit beat up. Do you want a new one?"],
- ["Not on your life! That little droid and I have been through a lot together.You okay, Artoo?"],
- ["Okay, easy she goes!"],
- ["Hang on tight, Artoo, you've got to come back."],
- ["You wouldn't want my life to get boring, would you?"],
- ["Luke, the Force will be with you."],
- ["Stand-by alert. Death Star approaching. Estimated time to firing range, fifteen minutes."],
- ["All wings report in."],
- ["Red Ten standing by."],
- ["Red Seven standing by."],
- ["Red Three standing by."],
- ["Red Six standing by."],
- ["Red Nine standing by."],
- ["Red Two standing by."],
- ["Red Eleven standing by."],
- ["Red Five standing by."],
- ["Lock S-foils in attack position."],
- ["We're passing through their magnetic field."],
- ["Hold tight!"],
- ["Switch your deflectors on."],
- ["Double front!"],
- ["Look at the size of that thing!"],
- ["Cut the chatter, Red Two."],
- ["Accelerate to attack speed. This is it, boys!"],
- ["Red Leader, this is Gold Leader."],
- ["I copy, Gold Leader."],
- ["We're starting for the target shaft now."],
- ["We're in position. I'm going to cut across the axis and try and draw their fire."],
- ["Heavy fire, boss! Twenty-threedegrees."],
- ["I see it. Stay low. "],
- ["This is Red Five! I'm going in!"],
- ["Luke, pull up!"],
- ["Are you all right?"],
- ["I got a little cooked, but I'm okay."],
- ["We count thirty Rebel ships, Lord Vader. But they're so small they're evading our turbo-lasers!"],
- ["We'll have to destroy them ship to ship. Get the crews to their fighters."],
- ["Luke, let me know when you're going in."],
- ["I'm on my way in now..."],
- ["Watch yourself! There's a lot of fire coming from the right side of that deflection tower."],
- ["I'm on it."],
- ["Squad leaders, we've picked up a new group of signals. Enemy fighters coming your way."],
- ["My scope's negative. I don't see anything."],
- ["Keep up your visual scanning. With all this jamming, they'll be on top of you before your scope can pick them up."],
- ["Biggs! You've picked one up... watch it!"],
- ["I can't see it! Where is he?!"],
- ["He's on me tight, I can't shake him... I can't shake him."],
- ["Hang on, Biggs, I'm coming in."],
- ["Got him!"],
- ["Several fighters have broken off from the main group. Come with me!"],
- ["Pull in! Luke... pull in!"],
- ["Watch your back, Luke!"],
- ["Watch your back! Fighter's above you, coming in!"],
- ["I'm hit, but not bad."],
- ["Artoo, see what you can do with it. Hang on back there."],
- ["Red Six..."],
- ["Can you see Red Five?"],
- ["There's a heavy fire zone on this side. Red Five, where are you?"],
- ["I can't shake him!"],
- ["I'm on him, Luke!"],
- ["Hold on!"],
- ["Blast it! Wedge where are you?"],
- ["Thanks, Wedge."],
- ["Good shooting, Wedge!"],
- ["Red Leader..."],
- ["... This is Gold Leader. We're starting out attack run."],
- ["I copy, Gold Leader. Move into position."],
- ["Stay in attack formation!"],
- ["The exhaust post is..."],
- ["... marked and locked in!"],
- ["Switch power to front deflection screens."],
- ["How many guns do you think, Gold Five?"],
- ["I'd say about twenty guns. Some on the surface, some on the towers."],
- ["Death Star will be in range in five minutes. "],
- ["Switch to targeting computer."],
- ["Computer's locked. Getting a signal."],
- ["The guns... they've stopped!"],
- ["Stabilize your rear deflectors. Watch for enemy fighters."],
- ["They've coming in! Three marks at two ten."],
- ["I'll take them myself! Cover me!"],
- ["Yes, sir."],
- ["I can't maneuver!"],
- ["Stay on target."],
- ["We're too close."],
- ["Stay on target!"],
- ["Loosen up!"],
- ["Gold Five to Red Leader..."],
- ["Lost Tiree, lost Dutch."],
- ["I copy, Gold Five."],
- ["They came from behind...."],
- ["We've analyzed their attack, sir, and there is a danger. Should I have your ship standing by?"],
- ["Evacuate? In out moment of triumph? I think you overestimate their chances!"],
- ["Rebel base, three minutes and closing."],
- ["Red Group, this is Red Leader."],
- ["Rendezvous at mark six point one."],
- ["This is Red Two. Flying towards you."],
- ["Red Three, standing by."],
- ["Red Leader, this is Base One. Keep half your group out of range for the next run."],
- ["Copy, Base One. Luke, take Red Two and Three. Hold up here and wait for my signal... to start your run."],
- ["This is it!"],
- ["We should be able to see it by now."],
- ["Keep your eyes open for those fighters!"],
- ["There's too much interference!"],
- ["Red Five, can you see them from where you are?"],
- ["No sign of any... wait!"],
- ["Coming in point three five."],
- ["I see them."],
- ["I'm in range."],
- ["Target's coming up!"],
- ["Just hold them off for a few seconds."],
- ["Close up formation."],
- ["Almost there!"],
- ["You'd better let her loose."],
- ["Almost there!"],
- ["I can't hold them!"],
- ["It's away!"],
- ["It's a hit!"],
- ["Negative."],
- ["Negative! It didn't go in, it just impacted on the surface."],
- ["Red Leader, we're right above you. Turn to point..."],
- ["... oh-five; we'll cover for you."],
- ["Stay there..."],
- ["... I just lost my starboard engine."],
- ["Get set up for your attack run."],
- ["Rebel base, one minute and closing."],
- ["Biggs, Wedge, let's close it up. We're going in. We're going in full throttle."],
- ["Right with you, boss."],
- ["Luke, at that speed will you be able to pull out in time?"],
- ["It'll be just like Beggar's Canyon back home."],
- ["We'll stay back far enough to cover you."],
- ["My scope shows the tower, but I can't see the exhaust port! Are you sure the computer can hit it?"],
- ["Watch yourself! Increase speed full throttle!"],
- ["What about that tower?"],
- ["You worry about those fighters! I'll worry about the tower!"],
- ["Artoo... that, that stabilizer's broken loose again! See if you can't lock it down!"],
- ["I'm hit! I can't stay with you."],
- ["Get clear, Wedge."],
- ["You can't do any more good back there!"],
- ["Sorry!"],
- ["Let him go! Stay on the leader!"],
- ["Hurry, Luke, they're coming in much faster this time. I can't hold them!"],
- ["Artoo, try and increase the power!"],
- ["Hurry up, Luke!"],
- ["Wait!"],
- ["Rebel base, thirty seconds and closing."],
- ["I'm on the leader."],
- ["Hang on, Artoo!"],
- ["Use the Force, Luke."],
- ["Let go, Luke."],
- ["The Force is strong with this one!"],
- ["Luke, trust me."],
- ["His computer's off. Luke, you switched off your targeting computer. What's wrong?"],
- ["Nothing. I'm all right."],
- ["I've lost Artoo!"],
- ["The Death Star has cleared the planet. The Death Star has cleared the planet."],
- ["Rebel base, in range."],
- ["You may fire when ready."],
- ["Commence primary ignition."],
- ["I have you now."],
- ["What?"],
- ["Yahoo!"],
- ["Look out!"],
- ["You're all clear, kid."],
- ["Now let's blow this thing and go home!"],
- ["Stand by to fire at Rebel base."],
- ["Standing by."],
- ["Great shot, kid. That was one in a million."],
- ["Remember, the Force will be with you... always."],
- ["Luke! Luke! Luke!"],
- ["Hey! Hey!"],
- ["I knew you'd come back! I just knew it!"],
- ["Well, I wasn't gonna let you get all the credit and take all the reward."],
- ["Hey, I knew there was more to you than money."],
- ["Oh, no!"],
- ["Oh, my! Artoo! Can you hear me? Say something!You can repair him, can't you?"],
- ["We'll get to work on him right away."],
- ["You must repair him! Sir, if any of my circuits or gears will help, I'll gladly donate them."],
- ["He'll be all right."]
-]
-from gradio import inputs
-from gradio.inputs import Textbox
-from gradio import outputs
-
-generator2 = gr.Interface.load("huggingface/EleutherAI/gpt-neo-2.7B")
-generator3 = gr.Interface.load("huggingface/EleutherAI/gpt-j-6B")
-generator1 = gr.Interface.load("huggingface/gpt2-large")
-
-#gr.Parallel(generator1, generator2, generator3, inputs=gr.inputs.Textbox(lines=6, label="Enter a sentence to get another sentence."),title=title, examples=examples).launch()
-
-def complete_with_gpt(text):
- # Use the last 50 characters of the text as context
- return text[:-50] + generator1(text[-50:])
-
-with gr.Blocks() as demo:
- textbox = gr.Textbox(placeholder="Type here and press enter...", lines=4)
- btn = gr.Button("Generate")
-
- btn.click(complete_with_gpt, textbox, textbox)
-
-demo.launch()
\ No newline at end of file
diff --git a/spaces/PeepDaSlan9/AutoGPT/autogpt/commands/times.py b/spaces/PeepDaSlan9/AutoGPT/autogpt/commands/times.py
deleted file mode 100644
index 3c9b8a4fc67a251c9e81a8c4a725cd1e25fcbebe..0000000000000000000000000000000000000000
--- a/spaces/PeepDaSlan9/AutoGPT/autogpt/commands/times.py
+++ /dev/null
@@ -1,10 +0,0 @@
-from datetime import datetime
-
-
-def get_datetime() -> str:
- """Return the current date and time
-
- Returns:
- str: The current date and time
- """
- return "Current date and time: " + datetime.now().strftime("%Y-%m-%d %H:%M:%S")
diff --git a/spaces/PeepDaSlan9/HuggingFaceH4-starchat-alpha/app.py b/spaces/PeepDaSlan9/HuggingFaceH4-starchat-alpha/app.py
deleted file mode 100644
index 6ec294f2c2bda44625ca5f0fd3c666f7be665216..0000000000000000000000000000000000000000
--- a/spaces/PeepDaSlan9/HuggingFaceH4-starchat-alpha/app.py
+++ /dev/null
@@ -1,3 +0,0 @@
-import gradio as gr
-
-gr.Interface.load("models/HuggingFaceH4/starchat-alpha").launch()
\ No newline at end of file
diff --git a/spaces/Plachta/VITS-Umamusume-voice-synthesizer/ONNXVITS_utils.py b/spaces/Plachta/VITS-Umamusume-voice-synthesizer/ONNXVITS_utils.py
deleted file mode 100644
index b634ce380421571e6e07fb45dd59717b3f63115c..0000000000000000000000000000000000000000
--- a/spaces/Plachta/VITS-Umamusume-voice-synthesizer/ONNXVITS_utils.py
+++ /dev/null
@@ -1,19 +0,0 @@
-import torch
-import numpy as np
-import random
-import onnxruntime as ort
-def set_random_seed(seed=0):
- ort.set_seed(seed)
- torch.manual_seed(seed)
- torch.cuda.manual_seed(seed)
- torch.backends.cudnn.deterministic = True
- random.seed(seed)
- np.random.seed(seed)
-
-def runonnx(model_path, **kwargs):
- ort_session = ort.InferenceSession(model_path)
- outputs = ort_session.run(
- None,
- kwargs
- )
- return outputs
\ No newline at end of file
diff --git a/spaces/Plachta/VITS-Umamusume-voice-synthesizer/text/english.py b/spaces/Plachta/VITS-Umamusume-voice-synthesizer/text/english.py
deleted file mode 100644
index 6817392ba8a9eb830351de89fb7afc5ad72f5e42..0000000000000000000000000000000000000000
--- a/spaces/Plachta/VITS-Umamusume-voice-synthesizer/text/english.py
+++ /dev/null
@@ -1,188 +0,0 @@
-""" from https://github.com/keithito/tacotron """
-
-'''
-Cleaners are transformations that run over the input text at both training and eval time.
-
-Cleaners can be selected by passing a comma-delimited list of cleaner names as the "cleaners"
-hyperparameter. Some cleaners are English-specific. You'll typically want to use:
- 1. "english_cleaners" for English text
- 2. "transliteration_cleaners" for non-English text that can be transliterated to ASCII using
- the Unidecode library (https://pypi.python.org/pypi/Unidecode)
- 3. "basic_cleaners" if you do not want to transliterate (in this case, you should also update
- the symbols in symbols.py to match your data).
-'''
-
-
-# Regular expression matching whitespace:
-
-
-import re
-import inflect
-from unidecode import unidecode
-import eng_to_ipa as ipa
-_inflect = inflect.engine()
-_comma_number_re = re.compile(r'([0-9][0-9\,]+[0-9])')
-_decimal_number_re = re.compile(r'([0-9]+\.[0-9]+)')
-_pounds_re = re.compile(r'£([0-9\,]*[0-9]+)')
-_dollars_re = re.compile(r'\$([0-9\.\,]*[0-9]+)')
-_ordinal_re = re.compile(r'[0-9]+(st|nd|rd|th)')
-_number_re = re.compile(r'[0-9]+')
-
-# List of (regular expression, replacement) pairs for abbreviations:
-_abbreviations = [(re.compile('\\b%s\\.' % x[0], re.IGNORECASE), x[1]) for x in [
- ('mrs', 'misess'),
- ('mr', 'mister'),
- ('dr', 'doctor'),
- ('st', 'saint'),
- ('co', 'company'),
- ('jr', 'junior'),
- ('maj', 'major'),
- ('gen', 'general'),
- ('drs', 'doctors'),
- ('rev', 'reverend'),
- ('lt', 'lieutenant'),
- ('hon', 'honorable'),
- ('sgt', 'sergeant'),
- ('capt', 'captain'),
- ('esq', 'esquire'),
- ('ltd', 'limited'),
- ('col', 'colonel'),
- ('ft', 'fort'),
-]]
-
-
-# List of (ipa, lazy ipa) pairs:
-_lazy_ipa = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('r', 'ɹ'),
- ('æ', 'e'),
- ('ɑ', 'a'),
- ('ɔ', 'o'),
- ('ð', 'z'),
- ('θ', 's'),
- ('ɛ', 'e'),
- ('ɪ', 'i'),
- ('ʊ', 'u'),
- ('ʒ', 'ʥ'),
- ('ʤ', 'ʥ'),
- ('ˈ', '↓'),
-]]
-
-# List of (ipa, lazy ipa2) pairs:
-_lazy_ipa2 = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('r', 'ɹ'),
- ('ð', 'z'),
- ('θ', 's'),
- ('ʒ', 'ʑ'),
- ('ʤ', 'dʑ'),
- ('ˈ', '↓'),
-]]
-
-# List of (ipa, ipa2) pairs
-_ipa_to_ipa2 = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('r', 'ɹ'),
- ('ʤ', 'dʒ'),
- ('ʧ', 'tʃ')
-]]
-
-
-def expand_abbreviations(text):
- for regex, replacement in _abbreviations:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def collapse_whitespace(text):
- return re.sub(r'\s+', ' ', text)
-
-
-def _remove_commas(m):
- return m.group(1).replace(',', '')
-
-
-def _expand_decimal_point(m):
- return m.group(1).replace('.', ' point ')
-
-
-def _expand_dollars(m):
- match = m.group(1)
- parts = match.split('.')
- if len(parts) > 2:
- return match + ' dollars' # Unexpected format
- dollars = int(parts[0]) if parts[0] else 0
- cents = int(parts[1]) if len(parts) > 1 and parts[1] else 0
- if dollars and cents:
- dollar_unit = 'dollar' if dollars == 1 else 'dollars'
- cent_unit = 'cent' if cents == 1 else 'cents'
- return '%s %s, %s %s' % (dollars, dollar_unit, cents, cent_unit)
- elif dollars:
- dollar_unit = 'dollar' if dollars == 1 else 'dollars'
- return '%s %s' % (dollars, dollar_unit)
- elif cents:
- cent_unit = 'cent' if cents == 1 else 'cents'
- return '%s %s' % (cents, cent_unit)
- else:
- return 'zero dollars'
-
-
-def _expand_ordinal(m):
- return _inflect.number_to_words(m.group(0))
-
-
-def _expand_number(m):
- num = int(m.group(0))
- if num > 1000 and num < 3000:
- if num == 2000:
- return 'two thousand'
- elif num > 2000 and num < 2010:
- return 'two thousand ' + _inflect.number_to_words(num % 100)
- elif num % 100 == 0:
- return _inflect.number_to_words(num // 100) + ' hundred'
- else:
- return _inflect.number_to_words(num, andword='', zero='oh', group=2).replace(', ', ' ')
- else:
- return _inflect.number_to_words(num, andword='')
-
-
-def normalize_numbers(text):
- text = re.sub(_comma_number_re, _remove_commas, text)
- text = re.sub(_pounds_re, r'\1 pounds', text)
- text = re.sub(_dollars_re, _expand_dollars, text)
- text = re.sub(_decimal_number_re, _expand_decimal_point, text)
- text = re.sub(_ordinal_re, _expand_ordinal, text)
- text = re.sub(_number_re, _expand_number, text)
- return text
-
-
-def mark_dark_l(text):
- return re.sub(r'l([^aeiouæɑɔəɛɪʊ ]*(?: |$))', lambda x: 'ɫ'+x.group(1), text)
-
-
-def english_to_ipa(text):
- text = unidecode(text).lower()
- text = expand_abbreviations(text)
- text = normalize_numbers(text)
- phonemes = ipa.convert(text)
- phonemes = collapse_whitespace(phonemes)
- return phonemes
-
-
-def english_to_lazy_ipa(text):
- text = english_to_ipa(text)
- for regex, replacement in _lazy_ipa:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def english_to_ipa2(text):
- text = english_to_ipa(text)
- text = mark_dark_l(text)
- for regex, replacement in _ipa_to_ipa2:
- text = re.sub(regex, replacement, text)
- return text.replace('...', '…')
-
-
-def english_to_lazy_ipa2(text):
- text = english_to_ipa(text)
- for regex, replacement in _lazy_ipa2:
- text = re.sub(regex, replacement, text)
- return text
diff --git a/spaces/Pranjal-y/data_scraping_analysis/data_analysis.py b/spaces/Pranjal-y/data_scraping_analysis/data_analysis.py
deleted file mode 100644
index 7475679b4ba8e6e57513d8fbc740c0da7247567a..0000000000000000000000000000000000000000
--- a/spaces/Pranjal-y/data_scraping_analysis/data_analysis.py
+++ /dev/null
@@ -1,256 +0,0 @@
-import streamlit as st
-import pandas as pd
-from pathlib import Path
-from xml.etree import ElementTree as ET
-import json
-import altair as alt
-import plotly.express as px
-import plotly.graph_objects as go
-import numpy as np
-import re
-
-
-def detect_file_format(file_path):
- extension = Path(file_path).suffix.lower()
- if extension == '.csv':
- return 'csv'
- elif extension == '.json':
- return 'json'
- elif extension == '.xml':
- return 'xml'
- else:
- return 'unsupported'
-
-def read_csv(file_path):
- return pd.read_csv(file_path)
-
-def read_json(file_path):
- with open(file_path, 'r') as f:
- return json.load(f)
-
-def read_xml(file_path):
- tree = ET.parse(file_path)
- root = tree.getroot()
- data = []
- for item in root:
- item_data = {}
- for child in item:
- item_data[child.tag] = child.text
- data.append(item_data)
- return pd.DataFrame(data)
-
-def data_analysis_page(file_path):
- st.title("Data Analysis")
-
- file_format = detect_file_format(file_path)
- st.write(f"Retrieved file is in {file_format.upper()} format")
-
- if file_format == 'csv':
- data = read_csv(file_path)
- elif file_format == 'json':
- data = read_json(file_path)
- elif file_format == 'xml':
- data = read_xml(file_path)
- else:
- st.warning("Unsupported file format")
-
- # Display the retrieved data
- st.write(f"
", unsafe_allow_html=True)
- data_types = data.dtypes
- st.write(data_types)
-
- # Initialize session state
- if 'converted_data' not in st.session_state:
- converted_data = data.copy()
-
- # Select column for analysis
- selected_column = st.selectbox("Select a column for analysis:", data.columns)
- if selected_column:
- try:
- numeric_data = data[selected_column].apply(pd.to_numeric, errors='coerce')
- numeric_data = numeric_data.dropna() # Remove NaN values
- average = numeric_data.mean()
- total_rows = len(data)
- user_engagement = numeric_data.sum()
-
- # Create columns for layout
- col1, col2, col3 = st.columns(3)
- with col1:
- st.markdown('
', unsafe_allow_html=True)
- st.write(f"
Average {selected_column}
", unsafe_allow_html=True)
- st.write(f"
{average:.2f}
", unsafe_allow_html=True)
- st.markdown('
', unsafe_allow_html=True)
-
- with col2:
- st.markdown('
', unsafe_allow_html=True)
- st.write(f"
Total Number of Rows
", unsafe_allow_html=True)
- st.write(f"
{total_rows}
", unsafe_allow_html=True)
- st.markdown('
', unsafe_allow_html=True)
-
- with col3:
- st.markdown('
', unsafe_allow_html=True)
- st.write(f"
User Engagement
", unsafe_allow_html=True)
- st.write(f"
Sum of {selected_column}: {user_engagement}
", unsafe_allow_html=True)
- st.markdown('
', unsafe_allow_html=True)
-
- except ValueError:
- st.warning(f"Selected column '{selected_column}' contains non-numeric values.")
-
- # Convert column data type if you wish to
- # Allow user to select a column for data conversion
- column_to_convert = st.selectbox("Select a column for data conversion:", data.columns)
-
- # Check if the selected column is already numeric
- if pd.api.types.is_numeric_dtype(converted_data[column_to_convert]):
- st.warning(f"Column '{column_to_convert}' is already numeric. Please select a different column.")
- else:
- # Provide a button to initiate data conversion
- if st.button("Convert to Numeric"):
- # Remove commas from values if the column is not numeric
- converted_data[column_to_convert] = converted_data[column_to_convert].str.replace(',', '')
-
- try:
- # Convert to numeric
- converted_data[column_to_convert] = pd.to_numeric(converted_data[column_to_convert])
- st.success(f"Converted '{column_to_convert}' to numeric.")
- except ValueError:
- st.warning(f"Column '{column_to_convert}' contains non-numeric values.")
-
- # Display updated data types
- updated_data_types = converted_data.dtypes
- st.write(updated_data_types)
- st.write(converted_data)
-
- # Histogram of selected column with tooltips using Altair
- # Histogram of selected column with tooltips using Altair
- st.markdown('
', unsafe_allow_html=True)
-
- # Density plot using Altair
- # Density plot using Altair
- st.markdown('
', unsafe_allow_html=True)
- st.write('## Density Plot')
-
- # Get a list of available columns for X-axis selection
- available_columns1 = data.columns.tolist()
-
- # Provide a unique key for the radio button
- selected_column1 = st.radio("Select X-axis Column:", available_columns1, key='density_radio')
-
- # Create the density plot based on the selected column
- density_chart = alt.Chart(converted_data).mark_area().encode(
- alt.X(f'{selected_column1}:Q', title=selected_column1), # Use the selected column here
- alt.Y('density:Q', title='Density'),
- alt.Tooltip([f'{selected_column1}:Q', 'density:Q'])
- ).transform_density(
- selected_column1, # Use the selected column here
- as_=[selected_column1, 'density']
- ).properties(
- width=600,
- height=400,
- title=f'Density Plot of {selected_column1}'
- )
- st.altair_chart(density_chart, use_container_width=True)
- st.markdown('
', unsafe_allow_html=True)
-
- # Double bar chart test
- # Create a double bar chart comparing procedure_price and cred_procedure_price for available data
- st.write('## Double Bar Chart: Comparison of Procedure Prices')
-
- # Get a list of available columns for X and Y axis selection
- available_columns = data.columns.tolist()
-
- # Create dropdowns for selecting X and Y axes
- selected_x_column = st.selectbox("Select X-axis Column:", available_columns)
- selected_y_column = st.selectbox("Select Y-axis Column:", available_columns)
-
- if not converted_data.empty:
- chart = alt.Chart(converted_data).mark_bar().encode(
- x=alt.X(f'{selected_x_column}:N', title='Disease'),
- y=alt.Y(f'{selected_y_column}:Q', title='Price (INR)', scale=alt.Scale(domain=(0, 7000000))),
- color=alt.Color('type_of_procedure:N', title='Type of Procedure',
- scale=alt.Scale(range=['blue', 'orange'])),
- tooltip=[f'{selected_x_column}:N', f'{selected_y_column}:Q', 'type_of_procedure:N']
- ).transform_fold(
- [selected_x_column, selected_y_column],
- as_=['type_of_procedure', 'price']
- ).properties(
- width=600,
- height=400,
- title='Comparison of Procedure Prices by Disease'
- )
- st.markdown('
- """
- st.sidebar.markdown(sidebar, unsafe_allow_html=True)
-
-def set_state_if_absent(key, value):
- if key not in st.session_state:
- st.session_state[key] = value
-
-def reset_results(*args):
- st.session_state.results = None
\ No newline at end of file
diff --git a/spaces/VikasKumar01/My_AI_chatbot/app.py b/spaces/VikasKumar01/My_AI_chatbot/app.py
deleted file mode 100644
index 33bdc69d00ecf6a7e38ca8d8c191f8ea75e2d94b..0000000000000000000000000000000000000000
--- a/spaces/VikasKumar01/My_AI_chatbot/app.py
+++ /dev/null
@@ -1,34 +0,0 @@
-import os
-import gradio as gr
-from langchain.chat_models import ChatOpenAI
-from langchain import LLMChain, PromptTemplate
-from langchain.memory import ConversationBufferMemory
-
-OPENAI_API_KEY=os.getenv('OPENAI_API_KEY')
-
-template = """Meet satvika, MY youthful and witty personal assistant! At 21 year old, she's full of energy and always egaer to help. satvika's goal is to assist you with any question or problems you might have. Her enthusiam shines through in every response, making interactions with her enjoyable and engaging.
-{chat_history}
-User: {user_message}
-Chatbot:"""
-
-prompt = PromptTemplate(
- input_variables=["chat_history", "user_message"], template=template
-)
-
-memory = ConversationBufferMemory(memory_key="chat_history")
-
-llm_chain = LLMChain(
- llm=ChatOpenAI(temperature='0.5', model_name="gpt-3.5-turbo"),
- prompt=prompt,
- verbose=True,
- memory=memory,
-)
-
-def get_text_response(user_message,history):
- response = llm_chain.predict(user_message = user_message)
- return response
-
-demo = gr.ChatInterface(get_text_response)
-
-if __name__ == "__main__":
- demo.launch() #To create a public link, set `share=True` in `launch()`. To enable errors and logs, set `debug=True` in `launch()`.
diff --git a/spaces/Wootang01/question_generator_two/README.md b/spaces/Wootang01/question_generator_two/README.md
deleted file mode 100644
index 6a24d36212902f6641872e9fd15001a391e462bb..0000000000000000000000000000000000000000
--- a/spaces/Wootang01/question_generator_two/README.md
+++ /dev/null
@@ -1,45 +0,0 @@
----
-title: Question_generator_two
-emoji: 📈
-colorFrom: yellow
-colorTo: purple
-sdk: gradio
-app_file: app.py
-pinned: false
----
-
-# Configuration
-
-`title`: _string_
-Display title for the Space
-
-`emoji`: _string_
-Space emoji (emoji-only character allowed)
-
-`colorFrom`: _string_
-Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
-
-`colorTo`: _string_
-Color for Thumbnail gradient (red, yellow, green, blue, indigo, purple, pink, gray)
-
-`sdk`: _string_
-Can be either `gradio`, `streamlit`, or `static`
-
-`sdk_version` : _string_
-Only applicable for `streamlit` SDK.
-See [doc](https://hf.co/docs/hub/spaces) for more info on supported versions.
-
-`app_file`: _string_
-Path to your main application file (which contains either `gradio` or `streamlit` Python code, or `static` html code).
-Path is relative to the root of the repository.
-
-`models`: _List[string]_
-HF model IDs (like "gpt2" or "deepset/roberta-base-squad2") used in the Space.
-Will be parsed automatically from your code if not specified here.
-
-`datasets`: _List[string]_
-HF dataset IDs (like "common_voice" or "oscar-corpus/OSCAR-2109") used in the Space.
-Will be parsed automatically from your code if not specified here.
-
-`pinned`: _boolean_
-Whether the Space stays on top of your list.
diff --git a/spaces/XzJosh/otto-Bert-VITS2/commons.py b/spaces/XzJosh/otto-Bert-VITS2/commons.py
deleted file mode 100644
index 9ad0444b61cbadaa388619986c2889c707d873ce..0000000000000000000000000000000000000000
--- a/spaces/XzJosh/otto-Bert-VITS2/commons.py
+++ /dev/null
@@ -1,161 +0,0 @@
-import math
-import numpy as np
-import torch
-from torch import nn
-from torch.nn import functional as F
-
-
-def init_weights(m, mean=0.0, std=0.01):
- classname = m.__class__.__name__
- if classname.find("Conv") != -1:
- m.weight.data.normal_(mean, std)
-
-
-def get_padding(kernel_size, dilation=1):
- return int((kernel_size*dilation - dilation)/2)
-
-
-def convert_pad_shape(pad_shape):
- l = pad_shape[::-1]
- pad_shape = [item for sublist in l for item in sublist]
- return pad_shape
-
-
-def intersperse(lst, item):
- result = [item] * (len(lst) * 2 + 1)
- result[1::2] = lst
- return result
-
-
-def kl_divergence(m_p, logs_p, m_q, logs_q):
- """KL(P||Q)"""
- kl = (logs_q - logs_p) - 0.5
- kl += 0.5 * (torch.exp(2. * logs_p) + ((m_p - m_q)**2)) * torch.exp(-2. * logs_q)
- return kl
-
-
-def rand_gumbel(shape):
- """Sample from the Gumbel distribution, protect from overflows."""
- uniform_samples = torch.rand(shape) * 0.99998 + 0.00001
- return -torch.log(-torch.log(uniform_samples))
-
-
-def rand_gumbel_like(x):
- g = rand_gumbel(x.size()).to(dtype=x.dtype, device=x.device)
- return g
-
-
-def slice_segments(x, ids_str, segment_size=4):
- ret = torch.zeros_like(x[:, :, :segment_size])
- for i in range(x.size(0)):
- idx_str = ids_str[i]
- idx_end = idx_str + segment_size
- ret[i] = x[i, :, idx_str:idx_end]
- return ret
-
-
-def rand_slice_segments(x, x_lengths=None, segment_size=4):
- b, d, t = x.size()
- if x_lengths is None:
- x_lengths = t
- ids_str_max = x_lengths - segment_size + 1
- ids_str = (torch.rand([b]).to(device=x.device) * ids_str_max).to(dtype=torch.long)
- ret = slice_segments(x, ids_str, segment_size)
- return ret, ids_str
-
-
-def get_timing_signal_1d(
- length, channels, min_timescale=1.0, max_timescale=1.0e4):
- position = torch.arange(length, dtype=torch.float)
- num_timescales = channels // 2
- log_timescale_increment = (
- math.log(float(max_timescale) / float(min_timescale)) /
- (num_timescales - 1))
- inv_timescales = min_timescale * torch.exp(
- torch.arange(num_timescales, dtype=torch.float) * -log_timescale_increment)
- scaled_time = position.unsqueeze(0) * inv_timescales.unsqueeze(1)
- signal = torch.cat([torch.sin(scaled_time), torch.cos(scaled_time)], 0)
- signal = F.pad(signal, [0, 0, 0, channels % 2])
- signal = signal.view(1, channels, length)
- return signal
-
-
-def add_timing_signal_1d(x, min_timescale=1.0, max_timescale=1.0e4):
- b, channels, length = x.size()
- signal = get_timing_signal_1d(length, channels, min_timescale, max_timescale)
- return x + signal.to(dtype=x.dtype, device=x.device)
-
-
-def cat_timing_signal_1d(x, min_timescale=1.0, max_timescale=1.0e4, axis=1):
- b, channels, length = x.size()
- signal = get_timing_signal_1d(length, channels, min_timescale, max_timescale)
- return torch.cat([x, signal.to(dtype=x.dtype, device=x.device)], axis)
-
-
-def subsequent_mask(length):
- mask = torch.tril(torch.ones(length, length)).unsqueeze(0).unsqueeze(0)
- return mask
-
-
-@torch.jit.script
-def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels):
- n_channels_int = n_channels[0]
- in_act = input_a + input_b
- t_act = torch.tanh(in_act[:, :n_channels_int, :])
- s_act = torch.sigmoid(in_act[:, n_channels_int:, :])
- acts = t_act * s_act
- return acts
-
-
-def convert_pad_shape(pad_shape):
- l = pad_shape[::-1]
- pad_shape = [item for sublist in l for item in sublist]
- return pad_shape
-
-
-def shift_1d(x):
- x = F.pad(x, convert_pad_shape([[0, 0], [0, 0], [1, 0]]))[:, :, :-1]
- return x
-
-
-def sequence_mask(length, max_length=None):
- if max_length is None:
- max_length = length.max()
- x = torch.arange(max_length, dtype=length.dtype, device=length.device)
- return x.unsqueeze(0) < length.unsqueeze(1)
-
-
-def generate_path(duration, mask):
- """
- duration: [b, 1, t_x]
- mask: [b, 1, t_y, t_x]
- """
- device = duration.device
-
- b, _, t_y, t_x = mask.shape
- cum_duration = torch.cumsum(duration, -1)
-
- cum_duration_flat = cum_duration.view(b * t_x)
- path = sequence_mask(cum_duration_flat, t_y).to(mask.dtype)
- path = path.view(b, t_x, t_y)
- path = path - F.pad(path, convert_pad_shape([[0, 0], [1, 0], [0, 0]]))[:, :-1]
- path = path.unsqueeze(1).transpose(2,3) * mask
- return path
-
-
-def clip_grad_value_(parameters, clip_value, norm_type=2):
- if isinstance(parameters, torch.Tensor):
- parameters = [parameters]
- parameters = list(filter(lambda p: p.grad is not None, parameters))
- norm_type = float(norm_type)
- if clip_value is not None:
- clip_value = float(clip_value)
-
- total_norm = 0
- for p in parameters:
- param_norm = p.grad.data.norm(norm_type)
- total_norm += param_norm.item() ** norm_type
- if clip_value is not None:
- p.grad.data.clamp_(min=-clip_value, max=clip_value)
- total_norm = total_norm ** (1. / norm_type)
- return total_norm
diff --git a/spaces/YUANAI/DiffspeechResearch/utils/metrics/dtw.py b/spaces/YUANAI/DiffspeechResearch/utils/metrics/dtw.py
deleted file mode 100644
index 829e8e160355f8729b8e478bc4a24ca8597df58e..0000000000000000000000000000000000000000
--- a/spaces/YUANAI/DiffspeechResearch/utils/metrics/dtw.py
+++ /dev/null
@@ -1,160 +0,0 @@
-from numpy import array, zeros, full, argmin, inf, ndim
-from scipy.spatial.distance import cdist
-from math import isinf
-
-
-def dtw(x, y, dist, warp=1, w=inf, s=1.0):
- """
- Computes Dynamic Time Warping (DTW) of two sequences.
-
- :param array x: N1*M array
- :param array y: N2*M array
- :param func dist: distance used as cost measure
- :param int warp: how many shifts are computed.
- :param int w: window size limiting the maximal distance between indices of matched entries |i,j|.
- :param float s: weight applied on off-diagonal moves of the path. As s gets larger, the warping path is increasingly biased towards the diagonal
- Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path.
- """
- assert len(x)
- assert len(y)
- assert isinf(w) or (w >= abs(len(x) - len(y)))
- assert s > 0
- r, c = len(x), len(y)
- if not isinf(w):
- D0 = full((r + 1, c + 1), inf)
- for i in range(1, r + 1):
- D0[i, max(1, i - w):min(c + 1, i + w + 1)] = 0
- D0[0, 0] = 0
- else:
- D0 = zeros((r + 1, c + 1))
- D0[0, 1:] = inf
- D0[1:, 0] = inf
- D1 = D0[1:, 1:] # view
- for i in range(r):
- for j in range(c):
- if (isinf(w) or (max(0, i - w) <= j <= min(c, i + w))):
- D1[i, j] = dist(x[i], y[j])
- C = D1.copy()
- jrange = range(c)
- for i in range(r):
- if not isinf(w):
- jrange = range(max(0, i - w), min(c, i + w + 1))
- for j in jrange:
- min_list = [D0[i, j]]
- for k in range(1, warp + 1):
- i_k = min(i + k, r)
- j_k = min(j + k, c)
- min_list += [D0[i_k, j] * s, D0[i, j_k] * s]
- D1[i, j] += min(min_list)
- if len(x) == 1:
- path = zeros(len(y)), range(len(y))
- elif len(y) == 1:
- path = range(len(x)), zeros(len(x))
- else:
- path = _traceback(D0)
- return D1[-1, -1], C, D1, path
-
-
-def accelerated_dtw(x, y, dist, warp=1):
- """
- Computes Dynamic Time Warping (DTW) of two sequences in a faster way.
- Instead of iterating through each element and calculating each distance,
- this uses the cdist function from scipy (https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.cdist.html)
-
- :param array x: N1*M array
- :param array y: N2*M array
- :param string or func dist: distance parameter for cdist. When string is given, cdist uses optimized functions for the distance metrics.
- If a string is passed, the distance function can be 'braycurtis', 'canberra', 'chebyshev', 'cityblock', 'correlation', 'cosine', 'dice', 'euclidean', 'hamming', 'jaccard', 'kulsinski', 'mahalanobis', 'matching', 'minkowski', 'rogerstanimoto', 'russellrao', 'seuclidean', 'sokalmichener', 'sokalsneath', 'sqeuclidean', 'wminkowski', 'yule'.
- :param int warp: how many shifts are computed.
- Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path.
- """
- assert len(x)
- assert len(y)
- if ndim(x) == 1:
- x = x.reshape(-1, 1)
- if ndim(y) == 1:
- y = y.reshape(-1, 1)
- r, c = len(x), len(y)
- D0 = zeros((r + 1, c + 1))
- D0[0, 1:] = inf
- D0[1:, 0] = inf
- D1 = D0[1:, 1:]
- D0[1:, 1:] = cdist(x, y, dist)
- C = D1.copy()
- for i in range(r):
- for j in range(c):
- min_list = [D0[i, j]]
- for k in range(1, warp + 1):
- min_list += [D0[min(i + k, r), j],
- D0[i, min(j + k, c)]]
- D1[i, j] += min(min_list)
- if len(x) == 1:
- path = zeros(len(y)), range(len(y))
- elif len(y) == 1:
- path = range(len(x)), zeros(len(x))
- else:
- path = _traceback(D0)
- return D1[-1, -1], C, D1, path
-
-
-def _traceback(D):
- i, j = array(D.shape) - 2
- p, q = [i], [j]
- while (i > 0) or (j > 0):
- tb = argmin((D[i, j], D[i, j + 1], D[i + 1, j]))
- if tb == 0:
- i -= 1
- j -= 1
- elif tb == 1:
- i -= 1
- else: # (tb == 2):
- j -= 1
- p.insert(0, i)
- q.insert(0, j)
- return array(p), array(q)
-
-
-if __name__ == '__main__':
- w = inf
- s = 1.0
- if 1: # 1-D numeric
- from sklearn.metrics.pairwise import manhattan_distances
-
- x = [0, 0, 1, 1, 2, 4, 2, 1, 2, 0]
- y = [1, 1, 1, 2, 2, 2, 2, 3, 2, 0]
- dist_fun = manhattan_distances
- w = 1
- # s = 1.2
- elif 0: # 2-D numeric
- from sklearn.metrics.pairwise import euclidean_distances
-
- x = [[0, 0], [0, 1], [1, 1], [1, 2], [2, 2], [4, 3], [2, 3], [1, 1], [2, 2], [0, 1]]
- y = [[1, 0], [1, 1], [1, 1], [2, 1], [4, 3], [4, 3], [2, 3], [3, 1], [1, 2], [1, 0]]
- dist_fun = euclidean_distances
- else: # 1-D list of strings
- from nltk.metrics.distance import edit_distance
-
- # x = ['we', 'shelled', 'clams', 'for', 'the', 'chowder']
- # y = ['class', 'too']
- x = ['i', 'soon', 'found', 'myself', 'muttering', 'to', 'the', 'walls']
- y = ['see', 'drown', 'himself']
- # x = 'we talked about the situation'.split()
- # y = 'we talked about the situation'.split()
- dist_fun = edit_distance
- dist, cost, acc, path = dtw(x, y, dist_fun, w=w, s=s)
-
- # Vizualize
- from matplotlib import pyplot as plt
-
- plt.imshow(cost.T, origin='lower', cmap=plt.cm.Reds, interpolation='nearest')
- plt.plot(path[0], path[1], '-o') # relation
- plt.xticks(range(len(x)), x)
- plt.yticks(range(len(y)), y)
- plt.xlabel('x')
- plt.ylabel('y')
- plt.axis('tight')
- if isinf(w):
- plt.title('Minimum distance: {}, slope weight: {}'.format(dist, s))
- else:
- plt.title('Minimum distance: {}, window widht: {}, slope weight: {}'.format(dist, w, s))
- plt.show()
diff --git a/spaces/YaTharThShaRma999/Testtrial1/README.md b/spaces/YaTharThShaRma999/Testtrial1/README.md
deleted file mode 100644
index 9f24912c0d72bab8e8cbd3e6b39c1ba2c9e2a488..0000000000000000000000000000000000000000
--- a/spaces/YaTharThShaRma999/Testtrial1/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: WizardLM
-emoji: 😁
-colorFrom: yellow
-colorTo: green
-sdk: gradio
-sdk_version: 3.35.2
-app_file: app.py
-pinned: false
-license: apache-2.0
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/Zengyf-CVer/Gradio_YOLOv5_Det_v4/model_download/yolov5_model_p5_n.sh b/spaces/Zengyf-CVer/Gradio_YOLOv5_Det_v4/model_download/yolov5_model_p5_n.sh
deleted file mode 100644
index 5fc6d093f4b92e1ad735f8b513d01d95f4d53d5c..0000000000000000000000000000000000000000
--- a/spaces/Zengyf-CVer/Gradio_YOLOv5_Det_v4/model_download/yolov5_model_p5_n.sh
+++ /dev/null
@@ -1,4 +0,0 @@
-cd ./yolov5
-
-# 下载YOLOv5模型
-wget -c -t 0 https://github.com/ultralytics/yolov5/releases/download/v6.1/yolov5n.pt
diff --git a/spaces/ZhangYuanhan/Bamboo_ViT-B16_demo/README.md b/spaces/ZhangYuanhan/Bamboo_ViT-B16_demo/README.md
deleted file mode 100644
index 74a1caa7498f2c89cde79a3f031ec6a77758e45f..0000000000000000000000000000000000000000
--- a/spaces/ZhangYuanhan/Bamboo_ViT-B16_demo/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Bamboo ViT-B16 Demo
-emoji: 🎋
-colorFrom: blue
-colorTo: blue
-sdk: gradio
-sdk_version: 3.0.17
-app_file: app.py
-pinned: false
-license: cc-by-4.0
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
\ No newline at end of file
diff --git a/spaces/Zulqrnain/NewsSummarizer/README.md b/spaces/Zulqrnain/NewsSummarizer/README.md
deleted file mode 100644
index fb41e0c8d1344ac64d60398569e2e178e002cd13..0000000000000000000000000000000000000000
--- a/spaces/Zulqrnain/NewsSummarizer/README.md
+++ /dev/null
@@ -1,12 +0,0 @@
----
-title: NewsSummarizer
-emoji: 🌖
-colorFrom: green
-colorTo: red
-sdk: gradio
-sdk_version: 3.17.0
-app_file: app.py
-pinned: false
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git "a/spaces/a-v-bely/spanish-task-generator/\320\222\321\205\320\276\320\264.py" "b/spaces/a-v-bely/spanish-task-generator/\320\222\321\205\320\276\320\264.py"
deleted file mode 100644
index fe3238dda91a8d18eb410d8ddde3439e0dfc6cd4..0000000000000000000000000000000000000000
--- "a/spaces/a-v-bely/spanish-task-generator/\320\222\321\205\320\276\320\264.py"
+++ /dev/null
@@ -1,35 +0,0 @@
-import warnings
-import streamlit as st
-from utilities.utils import is_valid_uuid
-from utilities_database.user_database_widgets import LogIn
-
-warnings.filterwarnings('ignore')
-st.header('Добро пожаловать!')
-st.subheader('Вы используете инструмент по автоматической генерации лексико-грамматических заданий по'
- ' испанскому языку!')
-st.write('**Зарегистрируйтесь или войдите в аккаунт**')
-__login__obj = LogIn(auth_token=st.secrets['COURIER_AUTH_TOKEN'],
- company_name=st.secrets['COMPANY_NAME'],
- width=200, height=200,
- logout_button_name='Выйти',
- hide_menu_bool=False,
- hide_footer_bool=False,
- lottie_url='https://assets2.lottiefiles.com/packages/lf20_jcikwtux.json')
-LOGGED_IN = __login__obj.build_login_ui()
-st.session_state['-LOGGED_IN-'] = False
-# Check for username in cookies
-if '-USER_NAME-' not in st.session_state:
- if __login__obj.cookies.get('__streamlit_login_signup_ui_username__'):
- if not is_valid_uuid(__login__obj.cookies['__streamlit_login_signup_ui_username__']):
- st.session_state['-USER_NAME-'] = __login__obj.cookies['__streamlit_login_signup_ui_username__']
- st.session_state['-LOGGED_IN_BOOL-'] = True
-
-if LOGGED_IN:
- st.session_state['-LOGGED_IN_BOOL-'] = True
- # st.session_state['-USER_NAME-'] =
- st.success('Можете переходить к следующим вкладкам!')
-
-st.markdown('*Автор-разработчик: А.В.Белый, кафедра математической лингвистики, филологический факультет СПбГУ,'
- ' 3 курс, бакалавриат, "Прикладная, компьютерная и математическая лингвистика (английский язык)"*'
- '\n\n*Научный руководитель: канд. филол. наук, доц. О.А.Митрофанова*')
-st.markdown('*E-mail: st087202@student.spbu.ru*')
diff --git a/spaces/aaronstaclara/towards-financial-inclusion/README.md b/spaces/aaronstaclara/towards-financial-inclusion/README.md
deleted file mode 100644
index 3f2816f34ad52689ea847da4a5764cdec3d9267b..0000000000000000000000000000000000000000
--- a/spaces/aaronstaclara/towards-financial-inclusion/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Towards Financial Inclusion
-emoji: 🐨
-colorFrom: gray
-colorTo: green
-sdk: streamlit
-sdk_version: 1.10.0
-app_file: app.py
-pinned: false
-license: afl-3.0
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/abhishek/sketch-to-image/annotator/uniformer/mmdet_null/models/detectors/kd_one_stage.py b/spaces/abhishek/sketch-to-image/annotator/uniformer/mmdet_null/models/detectors/kd_one_stage.py
deleted file mode 100644
index 671ec19015c87fefd065b84ae887147f90cc892b..0000000000000000000000000000000000000000
--- a/spaces/abhishek/sketch-to-image/annotator/uniformer/mmdet_null/models/detectors/kd_one_stage.py
+++ /dev/null
@@ -1,100 +0,0 @@
-import mmcv
-import torch
-from mmcv.runner import load_checkpoint
-
-from .. import build_detector
-from ..builder import DETECTORS
-from .single_stage import SingleStageDetector
-
-
-@DETECTORS.register_module()
-class KnowledgeDistillationSingleStageDetector(SingleStageDetector):
- r"""Implementation of `Distilling the Knowledge in a Neural Network.
- `_.
-
- Args:
- teacher_config (str | dict): Config file path
- or the config object of teacher model.
- teacher_ckpt (str, optional): Checkpoint path of teacher model.
- If left as None, the model will not load any weights.
- """
-
- def __init__(self,
- backbone,
- neck,
- bbox_head,
- teacher_config,
- teacher_ckpt=None,
- eval_teacher=True,
- train_cfg=None,
- test_cfg=None,
- pretrained=None):
- super().__init__(backbone, neck, bbox_head, train_cfg, test_cfg,
- pretrained)
- self.eval_teacher = eval_teacher
- # Build teacher model
- if isinstance(teacher_config, str):
- teacher_config = mmcv.Config.fromfile(teacher_config)
- self.teacher_model = build_detector(teacher_config['model'])
- if teacher_ckpt is not None:
- load_checkpoint(
- self.teacher_model, teacher_ckpt, map_location='cpu')
-
- def forward_train(self,
- img,
- img_metas,
- gt_bboxes,
- gt_labels,
- gt_bboxes_ignore=None):
- """
- Args:
- img (Tensor): Input images of shape (N, C, H, W).
- Typically these should be mean centered and std scaled.
- img_metas (list[dict]): A List of image info dict where each dict
- has: 'img_shape', 'scale_factor', 'flip', and may also contain
- 'filename', 'ori_shape', 'pad_shape', and 'img_norm_cfg'.
- For details on the values of these keys see
- :class:`mmdet.datasets.pipelines.Collect`.
- gt_bboxes (list[Tensor]): Each item are the truth boxes for each
- image in [tl_x, tl_y, br_x, br_y] format.
- gt_labels (list[Tensor]): Class indices corresponding to each box
- gt_bboxes_ignore (None | list[Tensor]): Specify which bounding
- boxes can be ignored when computing the loss.
- Returns:
- dict[str, Tensor]: A dictionary of loss components.
- """
- x = self.extract_feat(img)
- with torch.no_grad():
- teacher_x = self.teacher_model.extract_feat(img)
- out_teacher = self.teacher_model.bbox_head(teacher_x)
- losses = self.bbox_head.forward_train(x, out_teacher, img_metas,
- gt_bboxes, gt_labels,
- gt_bboxes_ignore)
- return losses
-
- def cuda(self, device=None):
- """Since teacher_model is registered as a plain object, it is necessary
- to put the teacher model to cuda when calling cuda function."""
- self.teacher_model.cuda(device=device)
- return super().cuda(device=device)
-
- def train(self, mode=True):
- """Set the same train mode for teacher and student model."""
- if self.eval_teacher:
- self.teacher_model.train(False)
- else:
- self.teacher_model.train(mode)
- super().train(mode)
-
- def __setattr__(self, name, value):
- """Set attribute, i.e. self.name = value
-
- This reloading prevent the teacher model from being registered as a
- nn.Module. The teacher module is registered as a plain object, so that
- the teacher parameters will not show up when calling
- ``self.parameters``, ``self.modules``, ``self.children`` methods.
- """
- if name == 'teacher_model':
- object.__setattr__(self, name, value)
- else:
- super().__setattr__(name, value)
diff --git a/spaces/abhishek/sketch-to-image/annotator/uniformer_base/mmseg/datasets/dataset_wrappers.py b/spaces/abhishek/sketch-to-image/annotator/uniformer_base/mmseg/datasets/dataset_wrappers.py
deleted file mode 100644
index 1a22501e0804e44e3350fb1f7bb95cd01fa14583..0000000000000000000000000000000000000000
--- a/spaces/abhishek/sketch-to-image/annotator/uniformer_base/mmseg/datasets/dataset_wrappers.py
+++ /dev/null
@@ -1,62 +0,0 @@
-'''
- * Copyright (c) 2023 Salesforce, Inc.
- * All rights reserved.
- * SPDX-License-Identifier: Apache License 2.0
- * For full license text, see LICENSE.txt file in the repo root or http://www.apache.org/licenses/
- * By Can Qin
- * Modified from ControlNet repo: https://github.com/lllyasviel/ControlNet
- * Copyright (c) 2023 Lvmin Zhang and Maneesh Agrawala
- * Modified from MMCV repo: From https://github.com/open-mmlab/mmcv
- * Copyright (c) OpenMMLab. All rights reserved.
-'''
-
-from torch.utils.data.dataset import ConcatDataset as _ConcatDataset
-
-from .builder import DATASETS
-
-
-@DATASETS.register_module()
-class ConcatDataset(_ConcatDataset):
- """A wrapper of concatenated dataset.
-
- Same as :obj:`torch.utils.data.dataset.ConcatDataset`, but
- concat the group flag for image aspect ratio.
-
- Args:
- datasets (list[:obj:`Dataset`]): A list of datasets.
- """
-
- def __init__(self, datasets):
- super(ConcatDataset, self).__init__(datasets)
- self.CLASSES = datasets[0].CLASSES
- self.PALETTE = datasets[0].PALETTE
-
-
-@DATASETS.register_module()
-class RepeatDataset(object):
- """A wrapper of repeated dataset.
-
- The length of repeated dataset will be `times` larger than the original
- dataset. This is useful when the data loading time is long but the dataset
- is small. Using RepeatDataset can reduce the data loading time between
- epochs.
-
- Args:
- dataset (:obj:`Dataset`): The dataset to be repeated.
- times (int): Repeat times.
- """
-
- def __init__(self, dataset, times):
- self.dataset = dataset
- self.times = times
- self.CLASSES = dataset.CLASSES
- self.PALETTE = dataset.PALETTE
- self._ori_len = len(self.dataset)
-
- def __getitem__(self, idx):
- """Get item from original dataset."""
- return self.dataset[idx % self._ori_len]
-
- def __len__(self):
- """The length is multiplied by ``times``"""
- return self.times * self._ori_len
diff --git a/spaces/abrar-lohia/text-2-character-anim/pyrender/.eggs/pyglet-2.0.5-py3.10.egg/pyglet/libs/win32/winkey.py b/spaces/abrar-lohia/text-2-character-anim/pyrender/.eggs/pyglet-2.0.5-py3.10.egg/pyglet/libs/win32/winkey.py
deleted file mode 100644
index 9205b25345a4cabcc2701f69728f33d3a91b8935..0000000000000000000000000000000000000000
--- a/spaces/abrar-lohia/text-2-character-anim/pyrender/.eggs/pyglet-2.0.5-py3.10.egg/pyglet/libs/win32/winkey.py
+++ /dev/null
@@ -1,197 +0,0 @@
-from pyglet.window import key
-from .constants import *
-
-keymap = {
- ord('A'): key.A,
- ord('B'): key.B,
- ord('C'): key.C,
- ord('D'): key.D,
- ord('E'): key.E,
- ord('F'): key.F,
- ord('G'): key.G,
- ord('H'): key.H,
- ord('I'): key.I,
- ord('J'): key.J,
- ord('K'): key.K,
- ord('L'): key.L,
- ord('M'): key.M,
- ord('N'): key.N,
- ord('O'): key.O,
- ord('P'): key.P,
- ord('Q'): key.Q,
- ord('R'): key.R,
- ord('S'): key.S,
- ord('T'): key.T,
- ord('U'): key.U,
- ord('V'): key.V,
- ord('W'): key.W,
- ord('X'): key.X,
- ord('Y'): key.Y,
- ord('Z'): key.Z,
- ord('0'): key._0,
- ord('1'): key._1,
- ord('2'): key._2,
- ord('3'): key._3,
- ord('4'): key._4,
- ord('5'): key._5,
- ord('6'): key._6,
- ord('7'): key._7,
- ord('8'): key._8,
- ord('9'): key._9,
- ord('\b'): key.BACKSPACE,
-
- # By experiment:
- 0x14: key.CAPSLOCK,
- 0x5d: key.MENU,
-
- # VK_LBUTTON: ,
- # VK_RBUTTON: ,
- VK_CANCEL: key.CANCEL,
- # VK_MBUTTON: ,
- # VK_BACK: ,
- VK_TAB: key.TAB,
- # VK_CLEAR: ,
- VK_RETURN: key.RETURN,
- VK_SHIFT: key.LSHIFT,
- VK_CONTROL: key.LCTRL,
- VK_MENU: key.LALT,
- VK_PAUSE: key.PAUSE,
- # VK_CAPITAL: ,
- # VK_KANA: ,
- # VK_HANGEUL: ,
- # VK_HANGUL: ,
- # VK_JUNJA: ,
- # VK_FINAL: ,
- # VK_HANJA: ,
- # VK_KANJI: ,
- VK_ESCAPE: key.ESCAPE,
- # VK_CONVERT: ,
- # VK_NONCONVERT: ,
- # VK_ACCEPT: ,
- # VK_MODECHANGE: ,
- VK_SPACE: key.SPACE,
- VK_PRIOR: key.PAGEUP,
- VK_NEXT: key.PAGEDOWN,
- VK_END: key.END,
- VK_HOME: key.HOME,
- VK_LEFT: key.LEFT,
- VK_UP: key.UP,
- VK_RIGHT: key.RIGHT,
- VK_DOWN: key.DOWN,
- # VK_SELECT: ,
- VK_PRINT: key.PRINT,
- # VK_EXECUTE: ,
- # VK_SNAPSHOT: ,
- VK_INSERT: key.INSERT,
- VK_DELETE: key.DELETE,
- VK_HELP: key.HELP,
- VK_LWIN: key.LWINDOWS,
- VK_RWIN: key.RWINDOWS,
- # VK_APPS: ,
- VK_NUMPAD0: key.NUM_0,
- VK_NUMPAD1: key.NUM_1,
- VK_NUMPAD2: key.NUM_2,
- VK_NUMPAD3: key.NUM_3,
- VK_NUMPAD4: key.NUM_4,
- VK_NUMPAD5: key.NUM_5,
- VK_NUMPAD6: key.NUM_6,
- VK_NUMPAD7: key.NUM_7,
- VK_NUMPAD8: key.NUM_8,
- VK_NUMPAD9: key.NUM_9,
- VK_MULTIPLY: key.NUM_MULTIPLY,
- VK_ADD: key.NUM_ADD,
- # VK_SEPARATOR: ,
- VK_SUBTRACT: key.NUM_SUBTRACT,
- VK_DECIMAL: key.NUM_DECIMAL,
- VK_DIVIDE: key.NUM_DIVIDE,
- VK_F1: key.F1,
- VK_F2: key.F2,
- VK_F3: key.F3,
- VK_F4: key.F4,
- VK_F5: key.F5,
- VK_F6: key.F6,
- VK_F7: key.F7,
- VK_F8: key.F8,
- VK_F9: key.F9,
- VK_F10: key.F10,
- VK_F11: key.F11,
- VK_F12: key.F12,
- VK_F13: key.F13,
- VK_F14: key.F14,
- VK_F15: key.F15,
- VK_F16: key.F16,
- VK_F17: key.F17,
- VK_F18: key.F18,
- VK_F19: key.F19,
- VK_F20: key.F20,
- VK_F21: key.F21,
- VK_F22: key.F22,
- VK_F23: key.F23,
- VK_F24: key.F24,
- VK_NUMLOCK: key.NUMLOCK,
- VK_SCROLL: key.SCROLLLOCK,
- VK_LSHIFT: key.LSHIFT,
- VK_RSHIFT: key.RSHIFT,
- VK_LCONTROL: key.LCTRL,
- VK_RCONTROL: key.RCTRL,
- VK_LMENU: key.LALT,
- VK_RMENU: key.RALT,
- # VK_PROCESSKEY: ,
- # VK_ATTN: ,
- # VK_CRSEL: ,
- # VK_EXSEL: ,
- # VK_EREOF: ,
- # VK_PLAY: ,
- # VK_ZOOM: ,
- # VK_NONAME: ,
- # VK_PA1: ,
- # VK_OEM_CLEAR: ,
- # VK_XBUTTON1: ,
- # VK_XBUTTON2: ,
- # VK_VOLUME_MUTE: ,
- # VK_VOLUME_DOWN: ,
- # VK_VOLUME_UP: ,
- # VK_MEDIA_NEXT_TRACK: ,
- # VK_MEDIA_PREV_TRACK: ,
- # VK_MEDIA_PLAY_PAUSE: ,
- # VK_BROWSER_BACK: ,
- # VK_BROWSER_FORWARD: ,
-}
-
-# Keys that must be translated via MapVirtualKey, as the virtual key code
-# is language and keyboard dependent.
-chmap = {
- ord('!'): key.EXCLAMATION,
- ord('"'): key.DOUBLEQUOTE,
- ord('#'): key.HASH,
- ord('$'): key.DOLLAR,
- ord('%'): key.PERCENT,
- ord('&'): key.AMPERSAND,
- ord("'"): key.APOSTROPHE,
- ord('('): key.PARENLEFT,
- ord(')'): key.PARENRIGHT,
- ord('*'): key.ASTERISK,
- ord('+'): key.PLUS,
- ord(','): key.COMMA,
- ord('-'): key.MINUS,
- ord('.'): key.PERIOD,
- ord('/'): key.SLASH,
- ord(':'): key.COLON,
- ord(';'): key.SEMICOLON,
- ord('<'): key.LESS,
- ord('='): key.EQUAL,
- ord('>'): key.GREATER,
- ord('?'): key.QUESTION,
- ord('@'): key.AT,
- ord('['): key.BRACKETLEFT,
- ord('\\'): key.BACKSLASH,
- ord(']'): key.BRACKETRIGHT,
- ord('\x5e'): key.ASCIICIRCUM,
- ord('_'): key.UNDERSCORE,
- ord('\x60'): key.GRAVE,
- ord('`'): key.QUOTELEFT,
- ord('{'): key.BRACELEFT,
- ord('|'): key.BAR,
- ord('}'): key.BRACERIGHT,
- ord('~'): key.ASCIITILDE,
-}
diff --git a/spaces/abrar-lohia/text-2-character-anim/pyrender/pyrender/constants.py b/spaces/abrar-lohia/text-2-character-anim/pyrender/pyrender/constants.py
deleted file mode 100644
index 8a5785b6fdb21910a174252c5af2f05b40ece4a5..0000000000000000000000000000000000000000
--- a/spaces/abrar-lohia/text-2-character-anim/pyrender/pyrender/constants.py
+++ /dev/null
@@ -1,149 +0,0 @@
-DEFAULT_Z_NEAR = 0.05 # Near clipping plane, in meters
-DEFAULT_Z_FAR = 100.0 # Far clipping plane, in meters
-DEFAULT_SCENE_SCALE = 2.0 # Default scene scale
-MAX_N_LIGHTS = 4 # Maximum number of lights of each type allowed
-TARGET_OPEN_GL_MAJOR = 4 # Target OpenGL Major Version
-TARGET_OPEN_GL_MINOR = 1 # Target OpenGL Minor Version
-MIN_OPEN_GL_MAJOR = 3 # Minimum OpenGL Major Version
-MIN_OPEN_GL_MINOR = 3 # Minimum OpenGL Minor Version
-FLOAT_SZ = 4 # Byte size of GL float32
-UINT_SZ = 4 # Byte size of GL uint32
-SHADOW_TEX_SZ = 2048 # Width and Height of Shadow Textures
-TEXT_PADDING = 20 # Width of padding for rendering text (px)
-
-
-# Flags for render type
-class RenderFlags(object):
- """Flags for rendering in the scene.
-
- Combine them with the bitwise or. For example,
-
- >>> flags = OFFSCREEN | SHADOWS_DIRECTIONAL | VERTEX_NORMALS
-
- would result in an offscreen render with directional shadows and
- vertex normals enabled.
- """
- NONE = 0
- """Normal PBR Render."""
- DEPTH_ONLY = 1
- """Only render the depth buffer."""
- OFFSCREEN = 2
- """Render offscreen and return the depth and (optionally) color buffers."""
- FLIP_WIREFRAME = 4
- """Invert the status of wireframe rendering for each mesh."""
- ALL_WIREFRAME = 8
- """Render all meshes as wireframes."""
- ALL_SOLID = 16
- """Render all meshes as solids."""
- SHADOWS_DIRECTIONAL = 32
- """Render shadows for directional lights."""
- SHADOWS_POINT = 64
- """Render shadows for point lights."""
- SHADOWS_SPOT = 128
- """Render shadows for spot lights."""
- SHADOWS_ALL = 32 | 64 | 128
- """Render shadows for all lights."""
- VERTEX_NORMALS = 256
- """Render vertex normals."""
- FACE_NORMALS = 512
- """Render face normals."""
- SKIP_CULL_FACES = 1024
- """Do not cull back faces."""
- RGBA = 2048
- """Render the color buffer with the alpha channel enabled."""
- FLAT = 4096
- """Render the color buffer flat, with no lighting computations."""
- SEG = 8192
-
-
-class TextAlign:
- """Text alignment options for captions.
-
- Only use one at a time.
- """
- CENTER = 0
- """Center the text by width and height."""
- CENTER_LEFT = 1
- """Center the text by height and left-align it."""
- CENTER_RIGHT = 2
- """Center the text by height and right-align it."""
- BOTTOM_LEFT = 3
- """Put the text in the bottom-left corner."""
- BOTTOM_RIGHT = 4
- """Put the text in the bottom-right corner."""
- BOTTOM_CENTER = 5
- """Center the text by width and fix it to the bottom."""
- TOP_LEFT = 6
- """Put the text in the top-left corner."""
- TOP_RIGHT = 7
- """Put the text in the top-right corner."""
- TOP_CENTER = 8
- """Center the text by width and fix it to the top."""
-
-
-class GLTF(object):
- """Options for GL objects."""
- NEAREST = 9728
- """Nearest neighbor interpolation."""
- LINEAR = 9729
- """Linear interpolation."""
- NEAREST_MIPMAP_NEAREST = 9984
- """Nearest mipmapping."""
- LINEAR_MIPMAP_NEAREST = 9985
- """Linear mipmapping."""
- NEAREST_MIPMAP_LINEAR = 9986
- """Nearest mipmapping."""
- LINEAR_MIPMAP_LINEAR = 9987
- """Linear mipmapping."""
- CLAMP_TO_EDGE = 33071
- """Clamp to the edge of the texture."""
- MIRRORED_REPEAT = 33648
- """Mirror the texture."""
- REPEAT = 10497
- """Repeat the texture."""
- POINTS = 0
- """Render as points."""
- LINES = 1
- """Render as lines."""
- LINE_LOOP = 2
- """Render as a line loop."""
- LINE_STRIP = 3
- """Render as a line strip."""
- TRIANGLES = 4
- """Render as triangles."""
- TRIANGLE_STRIP = 5
- """Render as a triangle strip."""
- TRIANGLE_FAN = 6
- """Render as a triangle fan."""
-
-
-class BufFlags(object):
- POSITION = 0
- NORMAL = 1
- TANGENT = 2
- TEXCOORD_0 = 4
- TEXCOORD_1 = 8
- COLOR_0 = 16
- JOINTS_0 = 32
- WEIGHTS_0 = 64
-
-
-class TexFlags(object):
- NONE = 0
- NORMAL = 1
- OCCLUSION = 2
- EMISSIVE = 4
- BASE_COLOR = 8
- METALLIC_ROUGHNESS = 16
- DIFFUSE = 32
- SPECULAR_GLOSSINESS = 64
-
-
-class ProgramFlags:
- NONE = 0
- USE_MATERIAL = 1
- VERTEX_NORMALS = 2
- FACE_NORMALS = 4
-
-
-__all__ = ['RenderFlags', 'TextAlign', 'GLTF']
diff --git a/spaces/adirik/stylemc-demo/encoder4editing/models/latent_codes_pool.py b/spaces/adirik/stylemc-demo/encoder4editing/models/latent_codes_pool.py
deleted file mode 100644
index 0281d4b5e80f8eb26e824fa35b4f908dcb6634e6..0000000000000000000000000000000000000000
--- a/spaces/adirik/stylemc-demo/encoder4editing/models/latent_codes_pool.py
+++ /dev/null
@@ -1,55 +0,0 @@
-import random
-import torch
-
-
-class LatentCodesPool:
- """This class implements latent codes buffer that stores previously generated w latent codes.
- This buffer enables us to update discriminators using a history of generated w's
- rather than the ones produced by the latest encoder.
- """
-
- def __init__(self, pool_size):
- """Initialize the ImagePool class
- Parameters:
- pool_size (int) -- the size of image buffer, if pool_size=0, no buffer will be created
- """
- self.pool_size = pool_size
- if self.pool_size > 0: # create an empty pool
- self.num_ws = 0
- self.ws = []
-
- def query(self, ws):
- """Return w's from the pool.
- Parameters:
- ws: the latest generated w's from the generator
- Returns w's from the buffer.
- By 50/100, the buffer will return input w's.
- By 50/100, the buffer will return w's previously stored in the buffer,
- and insert the current w's to the buffer.
- """
- if self.pool_size == 0: # if the buffer size is 0, do nothing
- return ws
- return_ws = []
- for w in ws: # ws.shape: (batch, 512) or (batch, n_latent, 512)
- # w = torch.unsqueeze(image.data, 0)
- if w.ndim == 2:
- i = random.randint(0, len(w) - 1) # apply a random latent index as a candidate
- w = w[i]
- self.handle_w(w, return_ws)
- return_ws = torch.stack(return_ws, 0) # collect all the images and return
- return return_ws
-
- def handle_w(self, w, return_ws):
- if self.num_ws < self.pool_size: # if the buffer is not full; keep inserting current codes to the buffer
- self.num_ws = self.num_ws + 1
- self.ws.append(w)
- return_ws.append(w)
- else:
- p = random.uniform(0, 1)
- if p > 0.5: # by 50% chance, the buffer will return a previously stored latent code, and insert the current code into the buffer
- random_id = random.randint(0, self.pool_size - 1) # randint is inclusive
- tmp = self.ws[random_id].clone()
- self.ws[random_id] = w
- return_ws.append(tmp)
- else: # by another 50% chance, the buffer will return the current image
- return_ws.append(w)
diff --git a/spaces/adyjay/andite-anything-v4.0/README.md b/spaces/adyjay/andite-anything-v4.0/README.md
deleted file mode 100644
index 7759f37f1031c968e6431941c4bdb09bc7648ab9..0000000000000000000000000000000000000000
--- a/spaces/adyjay/andite-anything-v4.0/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Andite Anything V4.0
-emoji: 💻
-colorFrom: pink
-colorTo: yellow
-sdk: gradio
-sdk_version: 3.16.1
-app_file: app.py
-pinned: false
-license: unknown
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/akhaliq/Music_Source_Separation/bytesep/callbacks/__init__.py b/spaces/akhaliq/Music_Source_Separation/bytesep/callbacks/__init__.py
deleted file mode 100644
index e70c6c2a4fa8fcabfdb78502907d431b07158edc..0000000000000000000000000000000000000000
--- a/spaces/akhaliq/Music_Source_Separation/bytesep/callbacks/__init__.py
+++ /dev/null
@@ -1,76 +0,0 @@
-from typing import List
-
-import pytorch_lightning as pl
-import torch.nn as nn
-
-
-def get_callbacks(
- task_name: str,
- config_yaml: str,
- workspace: str,
- checkpoints_dir: str,
- statistics_path: str,
- logger: pl.loggers.TensorBoardLogger,
- model: nn.Module,
- evaluate_device: str,
-) -> List[pl.Callback]:
- r"""Get callbacks of a task and config yaml file.
-
- Args:
- task_name: str
- config_yaml: str
- dataset_dir: str
- workspace: str, containing useful files such as audios for evaluation
- checkpoints_dir: str, directory to save checkpoints
- statistics_dir: str, directory to save statistics
- logger: pl.loggers.TensorBoardLogger
- model: nn.Module
- evaluate_device: str
-
- Return:
- callbacks: List[pl.Callback]
- """
- if task_name == 'musdb18':
-
- from bytesep.callbacks.musdb18 import get_musdb18_callbacks
-
- return get_musdb18_callbacks(
- config_yaml=config_yaml,
- workspace=workspace,
- checkpoints_dir=checkpoints_dir,
- statistics_path=statistics_path,
- logger=logger,
- model=model,
- evaluate_device=evaluate_device,
- )
-
- elif task_name == 'voicebank-demand':
-
- from bytesep.callbacks.voicebank_demand import get_voicebank_demand_callbacks
-
- return get_voicebank_demand_callbacks(
- config_yaml=config_yaml,
- workspace=workspace,
- checkpoints_dir=checkpoints_dir,
- statistics_path=statistics_path,
- logger=logger,
- model=model,
- evaluate_device=evaluate_device,
- )
-
- elif task_name in ['vctk-musdb18', 'violin-piano', 'piano-symphony']:
-
- from bytesep.callbacks.instruments_callbacks import get_instruments_callbacks
-
- return get_instruments_callbacks(
- config_yaml=config_yaml,
- workspace=workspace,
- checkpoints_dir=checkpoints_dir,
- statistics_path=statistics_path,
- logger=logger,
- model=model,
- evaluate_device=evaluate_device,
- )
-
- else:
- raise NotImplementedError
diff --git a/spaces/akhaliq/lama/saicinpainting/evaluation/utils.py b/spaces/akhaliq/lama/saicinpainting/evaluation/utils.py
deleted file mode 100644
index 6d7c15c9242ed8a9bc59fbb3b450cca394720bb8..0000000000000000000000000000000000000000
--- a/spaces/akhaliq/lama/saicinpainting/evaluation/utils.py
+++ /dev/null
@@ -1,28 +0,0 @@
-from enum import Enum
-
-import yaml
-from easydict import EasyDict as edict
-import torch.nn as nn
-import torch
-
-
-def load_yaml(path):
- with open(path, 'r') as f:
- return edict(yaml.safe_load(f))
-
-
-def move_to_device(obj, device):
- if isinstance(obj, nn.Module):
- return obj.to(device)
- if torch.is_tensor(obj):
- return obj.to(device)
- if isinstance(obj, (tuple, list)):
- return [move_to_device(el, device) for el in obj]
- if isinstance(obj, dict):
- return {name: move_to_device(val, device) for name, val in obj.items()}
- raise ValueError(f'Unexpected type {type(obj)}')
-
-
-class SmallMode(Enum):
- DROP = "drop"
- UPSCALE = "upscale"
diff --git a/spaces/alamin655/websurfx/public/static/colorschemes/solarized-dark.css b/spaces/alamin655/websurfx/public/static/colorschemes/solarized-dark.css
deleted file mode 100644
index 44494f9e57eb9e2a3f043ab1072474fcd922a0ea..0000000000000000000000000000000000000000
--- a/spaces/alamin655/websurfx/public/static/colorschemes/solarized-dark.css
+++ /dev/null
@@ -1,11 +0,0 @@
-:root {
- --background-color: #002b36;
- --foreground-color: #c9e0e6;
- --color-one: #073642;
- --color-two: #2AA198ff;
- --color-three: #2AA198ff;
- --color-four: #EEE8D5ff;
- --color-five: #268bd2;
- --color-six: #d33682;
- --color-seven: #fff;
-}
diff --git a/spaces/amielle/patent-summarizer/util/textproc.py b/spaces/amielle/patent-summarizer/util/textproc.py
deleted file mode 100644
index 4f6b4d09a0ce0351d58eb790ca55a0000a2f6af2..0000000000000000000000000000000000000000
--- a/spaces/amielle/patent-summarizer/util/textproc.py
+++ /dev/null
@@ -1,86 +0,0 @@
-import re
-import unicodedata
-import requests
-from bs4 import BeautifulSoup
-
-def retrieve_parsed_doc(patent_information, summaries_generated):
- try:
- language_config = "en"
- if "https" in patent_information:
- patent_code = patent_information.split("/")[4]
- else:
- patent_code = patent_information
- URL = f"https://patents.google.com/patent/{patent_code}/{language_config}"
- page = requests.get(URL)
-
- soup = BeautifulSoup(page.content, 'lxml')
-
- if "Abstract" in summaries_generated:
- abstract = clean_text(soup.find({"div":{"class":"abstract"}}).prettify())
- else:
- abstract = None
-
- if "Background" in summaries_generated:
- background = clean_text(soup.find_all(itemprop="description",
- itemscope="")[-1:][0].prettify())
- else:
- background = None
-
- if "Claims" in summaries_generated:
- claims = soup.find(itemprop="claims")
- main_claim = claims.find_all({"div":{"class":"claim"}})
- main_claims = main_claim[0].select("div[class=claim]")
- formatted_claims = set()
- for i in main_claims:
- formatted_claims.add(clean_text(i.prettify()))
- try:
- formatted_claims.remove('')
- except:
- pass
- claim_list = sorted(list(formatted_claims), key=len, reverse=True)
- else:
- claim_list = None
-
- return [abstract, background, claim_list]
- except Exception as e:
- print(f'[ERROR] {e}')
- return None
-
-
-def get_word_index(s, limit):
- try:
- words = re.findall(r'\s*\S+\s*', s)
- return sum(map(len, words[:limit])) + len(words[limit]) - len(words[limit].lstrip())
- except:
- l = len(s)
- chr_limit = 3500
- return l if l < chr_limit else chr_limit
-
-
-def post_process(s):
- # Basic post-processing
-
- if s[0] == " ": s = s[1:]
- s = s.replace("- ", "-").replace(" .", ".")
- return ".".join(s.split(".")[:-1])+"."
-
-
-def clean_text(text):
- # TODO: optimize text cleaning
- reg = re.compile(r'<.*?>')
- cleaned = reg.sub('', text)
- cleaned = re.sub(r'\([^)]*\)', '', cleaned)
- cleaned = re.sub(r"(\w)([A-Z]+)", r'.', cleaned)
- cleaned = cleaned.strip()
- cleaned = cleaned.lstrip()
- cleaned = "".join(ch for ch in cleaned if unicodedata.category(ch)[0]!="C")
- cleaned = re.sub(' +', ' ', cleaned)
- cleaned = cleaned.replace(";", ", and")
- cleaned = cleaned.replace(":", "")
- cleaned = cleaned.replace(" .", ".")
- cleaned = cleaned.replace(" ,", ",")
- cleaned = cleaned.replace("\xa0", " ")
- cleaned = cleaned.lstrip('0123456789.- ') # remove nums at start
- cleaned = re.sub(r'\b(\w+)( \1\b)+', r'\1', cleaned) #remove repeated consecutive words
-
- return cleaned
\ No newline at end of file
diff --git a/spaces/anderbogia/dtp-asr-demo-v2/app.py b/spaces/anderbogia/dtp-asr-demo-v2/app.py
deleted file mode 100644
index c92c99b5433c2699720682a957e526f1c8723e01..0000000000000000000000000000000000000000
--- a/spaces/anderbogia/dtp-asr-demo-v2/app.py
+++ /dev/null
@@ -1,93 +0,0 @@
-import os
-#os.system("curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y") #Installing Rust manually
-#os.system("exec bash")
-#os.system("pip install --upgrade pip")
-os.system("pip install transformers==4.30.2") #Some interoperability issue with Wav2Vec2CTCTokenizer. Refer here: https://github.com/huggingface/transformers/pull/26349
-os.system("pip install tokenizers fairseq")
-os.system("pip install numpy==1.23.0") #NumPy 1.24 or less needed by Numba. Use 1.23, librosa still uses np.complex which was dropped in NumPy 1.24
-#os.system("pip install git+https://github.com/huggingface/transformers datasets[torch]")
-os.system("pip install torch accelerate torchaudio datasets librosa easymms")
-
-
-import gradio as gr
-from transformers import pipeline, Wav2Vec2ForCTC, AutoProcessor
-from datasets import load_dataset, Audio, Dataset
-import torch
-import librosa #For converting audio sample rate to 16k
-from easymms.models.tts import TTSModel #For TTS inference using EasyMMS
-
-LANG = "dtp" #Change to tih for Timugon Murut or iba for Iban
-model_id = "facebook/mms-1b-all"
-
-processor = AutoProcessor.from_pretrained(model_id)
-model = Wav2Vec2ForCTC.from_pretrained(model_id).to("cpu")
-processor.tokenizer.set_target_lang(LANG)
-model.load_adapter(LANG)
-
-asr_pipeline = pipeline(task = "automatic-speech-recognition", model = model_id) #Function that returns a dict, transcription stored in item with key "text"
-
-def preprocess(input): #Sets recording sampling rate to 16k and returns numpy ndarray from audio
- speech, sample_rate = librosa.load(input)
- speech = librosa.resample(speech, orig_sr=sample_rate, target_sr=16000)
- loaded_audio = Dataset.from_dict({"audio": [input]}).cast_column("audio", Audio(sampling_rate=16000))
- audio_to_array = loaded_audio[0]["audio"]["array"]
- return audio_to_array
-
-def run(input):
- inputs = processor(input, sampling_rate=16_000, return_tensors="pt")
- with torch.no_grad():
- outputs = model(**inputs).logits
- ids = torch.argmax(outputs, dim=-1)[0]
- transcription = processor.decode(ids)
- return transcription
-
-def transcribe(input): #Gradio UI wrapper function
- audioarray = preprocess(input) #Call preprocessor function
- out = run(audioarray)
- return out
-
-with gr.Blocks(theme = gr.themes.Soft()) as demo:
- gr.HTML(
- """
-
Ponutun Tuturan om Pomorolou Sinuat Boros Dusun
-
Poomitanan kopogunaan do somit tutun tuturan om pomorolou sinuat (speech recognition and text-to-speech models)
- pinoluda' di Woyotanud Tuturan Gumukabang Tagayo di Meta (Meta Massive Multilingual Speech Project)
- """)
-
- tts = TTSModel(LANG)
-
- def fn2(input):
- res = tts.synthesize(input)
- flip_tuple = (res[1], res[0]) #EasyMMS synthesize() returns Tuple(data, sample_rate) where data is a numpy.array and sample_rate is int,
- #but Gradio Audio() expects the same tuple but with the elements flipped
- return flip_tuple
-
- with gr.Row():
- with gr.Column(scale = 1):
- gr.HTML("""
""")
-
- gr.Markdown("""
- **Huminodun, nulai di somit pongulai kikito DALL-E**
-
- *Huminodun, generated by the image generation model DALL-E*
- """)
- with gr.Column(scale = 4):
- with gr.Tab("Rolou kumaa ginarit"):
- input_audio = gr.Audio(source = "microphone", type = "filepath", label = "Gakamai rolou nu")
- output_text = gr.components.Textbox(label = "Dalinsuat")
- button1 = gr.Button("Dalinsuato' | Transcribe")
- button1.click(transcribe, inputs = input_audio, outputs = output_text)
-
- with gr.Tab("Ginarit kumaa rolou"):
- input_text = gr.components.Textbox(label = "Ginarit", placeholder = "Potutakai suat nu hiti")
- button2 = gr.Button("Poulayo'")
- output_audio = gr.components.Audio(label = "Rolou pinoulai")
- button2.click(fn2, inputs = input_text, outputs = output_audio)
-
-demo.launch(debug = True)
\ No newline at end of file
diff --git a/spaces/andromeda123/captionscraft/app.py b/spaces/andromeda123/captionscraft/app.py
deleted file mode 100644
index 642d772f10985266fb8c3ffc02f3ee3548cd91a9..0000000000000000000000000000000000000000
--- a/spaces/andromeda123/captionscraft/app.py
+++ /dev/null
@@ -1,158 +0,0 @@
-# -*- coding: utf-8 -*-
-
-# !apt install imagemagick
-
-# !cat /etc/ImageMagick-6/policy.xml | sed 's/none/read,write/g'> /etc/ImageMagick-6/policy.xml
-
-# Place files in this path or modify the paths to point to where the files are
-srtfilename = "subtitles.txt"
-mp4filename = "video.mp4"
-
-import sys
-import os
-import subprocess
-import streamlit as st
-from faster_whisper import WhisperModel
-import time
-
-def save_uploadedfile(uploadedfile):
- with open(filename,"wb") as f:
- f.write(uploadedfile.getbuffer())
-
-def time_to_seconds(time_obj):
- return time_obj.hours * 3600 + time_obj.minutes * 60 + time_obj.seconds + time_obj.milliseconds / 1000
-
-
-def video2mp3(video_file, output_ext="mp3"):
- filename, ext = os.path.splitext(video_file)
- subprocess.call(["ffmpeg", "-y", "-i", video_file, f"{filename}.{output_ext}"],
- stdout=subprocess.DEVNULL,
- stderr=subprocess.STDOUT)
- return f"{filename}.{output_ext}"
-
-def translate(audio , model):
- options = dict(beam_size=5, best_of=5)
- translate_options = dict(task="translate", **options)
- result,info = model.transcribe(audio_file,**translate_options)
- return result
-
-def format_timestamp(time):
- if(time< 0): return "timestamp cannot be negative"
- time_in_ms = round(time* 1000.0)
-
- hours = time_in_ms // 3_600_000
- time_in_ms -= hours * 3_600_000
-
- minutes = time_in_ms // 60_000
- time_in_ms -= minutes * 60_000
-
- seconds = time_in_ms // 1_000
- time_in_ms -= seconds * 1_000
-
- return f"{hours}:{minutes:02d}:{seconds:02d},{time_in_ms:03d}"
-
-def write_srt(segments,filename):
- index=1
- file1 = open(filename, "w") # append mode
-
- for segment in segments:
- file1.write( f"{index}\n"
- f"{format_timestamp(segment.start)} --> "
- f"{format_timestamp(segment.end)}\n"
- f"{segment.text.strip().replace('-->', '->')}\n\n",)
- index+=1
-
-
-
-#############
-#PAGE SET UP
-#############
-
-st.set_page_config(page_title="CaptionsCraft",
- page_icon=":pen:",
- layout="wide",
- initial_sidebar_state="expanded"
- )
-
-
-#########
-#SIDEBAR
-########
-
-st.sidebar.header('Navigate to:')
-nav = st.sidebar.radio('',['Go to homepage', 'Generate subtitles'])
-st.sidebar.write('')
-st.sidebar.write('')
-st.sidebar.write('')
-st.sidebar.write('')
-st.sidebar.write('')
-
-
-#HOME
-#####
-
-if nav == 'Go to homepage':
-
- st.markdown("
CaptionCraft
", unsafe_allow_html=True)
- st.markdown("
🎬✎
", unsafe_allow_html=True)
- st.markdown("
Utilizing advanced Whisper AI, it effortlessly converts any language spoken in a video into accurate English subtitles. Bridging communication gaps seamlessly.
", unsafe_allow_html=True)
- st.write("""This app harnesses the cutting-edge power of the Whisper model to provide you with an unparalleled video subtitle generation experience.
-
-\n\nImagine watching a video in a language you don't understand, but with our app, you won't miss a single detail. Whether it's a captivating foreign short film, an informative documentary, or a heartwarming vlog, our app steps in to bridge the linguistic gap.
-
-\n\nPowered by Whisper AI, our app listens to the spoken words in the video and expertly converts them into accurate and contextually relevant English subtitles. It's like having your own personal interpreter working in real-time, enabling you to enjoy content from around the world without missing out on any crucial information.""")
-
- st.markdown("
How to use the app?
", unsafe_allow_html=True)
- st.write("""1) Navigate to the 'Generate subtitles' page using navigation bar on the left , and upload the video file.
- \n\n 2) Choose the whisper model size \n\n 3) Upload your file (limit is 500 mb) \n\n 4) Your subtitles.txt file will be downloaded
- \n\n 5) Using the file , subtitles can be imposed on any video using any standard video player application.""")
- st.write("Here is the repo link : [GitHub](https://github.com/s0ur-oranges/subtitle_generator)")
-
-
-if nav == 'Generate subtitles':
- filename="videofile"
-
- print("hello")
- st.write("Choose a model size from the following: ")
-
- model_size= st.radio("Model sizes",["no model selected","tiny","base","small","medium","large-v2"] , index=0 , label_visibility='hidden')
-
- st.write("")
-
- if model_size=="no model selected":
- st.write("Select a model size to continue")
-
-
- else:
- uploaded_file = st.file_uploader("Upload your file here...")
- # or run on CPU with INT8
- model = WhisperModel(model_size, device="cpu", compute_type="int8")
-
- if uploaded_file:
- save_uploadedfile(uploaded_file)
- print('file saved')
- st.write("Please wait while your video is getting processed.")
-
- input_video = filename
- audio_file = video2mp3(input_video)
-
- result = translate(audio_file,model)
-
- print('audio translated')
- subtitle_filename='subtitles.txt'
- write_srt(result,subtitle_filename)
-
- print('subtitle generated')
-
- with open(subtitle_filename, "rb") as file:
- btn = st.download_button(
-
- label="Download file",
- data=file,
- file_name="subtitles.txt"
- )
- st.write("Note: If you want to try another model size , reload the page before repeating the selection process.")
\ No newline at end of file
diff --git a/spaces/anshu-man853/webscrapping/app.py b/spaces/anshu-man853/webscrapping/app.py
deleted file mode 100644
index 4edc57d0cced41b08ca07f7e9becf9ce8c45ae2c..0000000000000000000000000000000000000000
--- a/spaces/anshu-man853/webscrapping/app.py
+++ /dev/null
@@ -1,31 +0,0 @@
-import gradio as gr
-import requests
-from bs4 import BeautifulSoup
-import re
-import html
-
-# Define the web scraping function
-def scrape_website(url):
- # Send a GET request to the website
- response = requests.get(url)
- html_content = response.content
- # Parse the HTML content using BeautifulSoup
- soup = BeautifulSoup(html_content, "html.parser")
- # Extract all text from the HTML
- text = soup.get_text()
- # Clean the text by removing extra whitespaces and special characters
- cleaned_text = re.sub(r"\s+", " ", text)
- cleaned_text = html.unescape(cleaned_text)
- return cleaned_text
-
-# Create a Gradio interface
-iface = gr.Interface(
- fn=scrape_website,
- inputs="text",
- outputs="text",
- title="Web Scraping",
- description="Enter a website URL to scrape its text",
- example="https://www.example.com"
-)
-
-iface.launch()
diff --git a/spaces/antonovmaxim/text-generation-webui-space/extensions/multimodal/multimodal_embedder.py b/spaces/antonovmaxim/text-generation-webui-space/extensions/multimodal/multimodal_embedder.py
deleted file mode 100644
index 62e99ca7c950bcdae65049d0cb426a2fa53ba2b7..0000000000000000000000000000000000000000
--- a/spaces/antonovmaxim/text-generation-webui-space/extensions/multimodal/multimodal_embedder.py
+++ /dev/null
@@ -1,178 +0,0 @@
-import base64
-import logging
-import re
-from dataclasses import dataclass
-from io import BytesIO
-from typing import Any, List, Optional
-
-import torch
-from PIL import Image
-
-from extensions.multimodal.pipeline_loader import load_pipeline
-from modules import shared
-from modules.text_generation import encode, get_max_prompt_length
-
-
-@dataclass
-class PromptPart:
- text: str
- image: Optional[Image.Image] = None
- is_image: bool = False
- input_ids: Optional[torch.Tensor] = None
- embedding: Optional[torch.Tensor] = None
-
-
-class MultimodalEmbedder:
- def __init__(self, params: dict):
- pipeline, source = load_pipeline(params)
- self.pipeline = pipeline
- logging.info(f'Multimodal: loaded pipeline {self.pipeline.name()} from pipelines/{source} ({self.pipeline.__class__.__name__})')
-
- def _split_prompt(self, prompt: str, load_images: bool = False) -> List[PromptPart]:
- """Splits a prompt into a list of `PromptParts` to separate image data from text.
- It will also append `image_start` and `image_end` before and after the image, and optionally parse and load the images,
- if `load_images` is `True`.
- """
- parts: List[PromptPart] = []
- curr = 0
- while True:
- match = re.search(r'', prompt[curr:])
- if match is None:
- # no more image tokens, append the rest of the prompt
- if curr > 0:
- # add image end token after last image
- parts.append(PromptPart(text=self.pipeline.image_end() + prompt[curr:]))
- else:
- parts.append(PromptPart(text=prompt))
- break
- # found an image, append image start token to the text
- if match.start() > 0:
- parts.append(PromptPart(text=prompt[curr:curr + match.start()] + self.pipeline.image_start()))
- else:
- parts.append(PromptPart(text=self.pipeline.image_start()))
- # append the image
- parts.append(PromptPart(
- text=match.group(0),
- image=Image.open(BytesIO(base64.b64decode(match.group(1)))) if load_images else None,
- is_image=True
- ))
- curr += match.end()
- return parts
-
- def _len_in_tokens_prompt_parts(self, parts: List[PromptPart]) -> int:
- """Total length in tokens of all `parts`"""
- tokens = 0
- for part in parts:
- if part.is_image:
- tokens += self.pipeline.num_image_embeds()
- elif part.input_ids is not None:
- tokens += len(part.input_ids)
- else:
- tokens += len(encode(part.text)[0])
- return tokens
-
- def len_in_tokens(self, prompt: str) -> int:
- """Total length in tokens for a given text `prompt`"""
- parts = self._split_prompt(prompt, False)
- return self._len_in_tokens_prompt_parts(parts)
-
- def _encode_single_text(self, part: PromptPart, add_bos_token: bool) -> PromptPart:
- """Encode a single prompt `part` to `input_ids`. Returns a `PromptPart`"""
- if part.is_image:
- placeholders = torch.ones((self.pipeline.num_image_embeds())) * self.pipeline.placeholder_token_id()
- part.input_ids = placeholders.to(shared.model.device, dtype=torch.int64)
- else:
- part.input_ids = encode(part.text, add_bos_token=add_bos_token)[0].to(shared.model.device, dtype=torch.int64)
- return part
-
- @staticmethod
- def _num_images(parts: List[PromptPart]) -> int:
- count = 0
- for part in parts:
- if part.is_image:
- count += 1
- return count
-
- def _encode_text(self, state, parts: List[PromptPart]) -> List[PromptPart]:
- """Encode text to token_ids, also truncate the prompt, if necessary.
-
- The chat/instruct mode should make prompts that fit in get_max_prompt_length, but if max_new_tokens are set
- such that the context + min_rows don't fit, we can get a prompt which is too long.
- We can't truncate image embeddings, as it leads to broken generation, so remove the images instead and warn the user
- """
- encoded: List[PromptPart] = []
- for i, part in enumerate(parts):
- encoded.append(self._encode_single_text(part, i == 0 and state['add_bos_token']))
-
- # truncation:
- max_len = get_max_prompt_length(state)
- removed_images = 0
-
- # 1. remove entire text/image blocks
- while self._len_in_tokens_prompt_parts(encoded[1:]) > max_len:
- if encoded[0].is_image:
- removed_images += 1
- encoded = encoded[1:]
-
- # 2. check if the last prompt part doesn't need to get truncated
- if self._len_in_tokens_prompt_parts(encoded) > max_len:
- if encoded[0].is_image:
- # don't truncate image embeddings, just remove the image, otherwise generation will be broken
- removed_images += 1
- encoded = encoded[1:]
- elif len(encoded) > 1 and encoded[0].text.endswith(self.pipeline.image_start()):
- # see if we can keep image_start token
- len_image_start = len(encode(self.pipeline.image_start(), add_bos_token=state['add_bos_token'])[0])
- if self._len_in_tokens_prompt_parts(encoded[1:]) + len_image_start > max_len:
- # we can't -> remove this text, and the image
- encoded = encoded[2:]
- removed_images += 1
- else:
- # we can -> just truncate the text
- trunc_len = self._len_in_tokens_prompt_parts(encoded) - max_len
- encoded[0].input_ids = encoded[0].input_ids[trunc_len:]
- elif len(encoded) > 0:
- # only one text left, truncate it normally
- trunc_len = self._len_in_tokens_prompt_parts(encoded) - max_len
- encoded[0].input_ids = encoded[0].input_ids[trunc_len:]
-
- # notify user if we truncated an image
- if removed_images > 0:
- logging.warning(f"Multimodal: removed {removed_images} image(s) from prompt. Try decreasing max_new_tokens if generation is broken")
-
- return encoded
-
- def _embed(self, parts: List[PromptPart]) -> List[PromptPart]:
- # batch images
- image_indicies = [i for i, part in enumerate(parts) if part.is_image]
- embedded = self.pipeline.embed_images([parts[i].image for i in image_indicies])
- for i, embeds in zip(image_indicies, embedded):
- parts[i].embedding = embeds
- # embed text
- for (i, part) in enumerate(parts):
- if not part.is_image:
- parts[i].embedding = self.pipeline.embed_tokens(part.input_ids)
- return parts
-
- def _remove_old_images(self, parts: List[PromptPart], params: dict) -> List[PromptPart]:
- if params['add_all_images_to_prompt']:
- return parts
- already_added = False
- for i, part in reversed(list(enumerate(parts))):
- if part.is_image:
- if already_added:
- parts[i].embedding = self.pipeline.placeholder_embeddings()
- else:
- already_added = True
- return parts
-
- def forward(self, prompt: str, state: Any, params: dict):
- prompt_parts = self._split_prompt(prompt, True)
- prompt_parts = self._encode_text(state, prompt_parts)
- prompt_parts = self._embed(prompt_parts)
- prompt_parts = self._remove_old_images(prompt_parts, params)
- embeds = tuple(part.embedding for part in prompt_parts)
- ids = tuple(part.input_ids for part in prompt_parts)
- input_embeds = torch.cat(embeds, dim=0)
- input_ids = torch.cat(ids, dim=0)
- return prompt, input_ids, input_embeds, self._num_images(prompt_parts)
diff --git a/spaces/aodianyun/stable-diffusion-webui/extensions/deforum/scripts/deforum_helpers/src/adabins/unet_adaptive_bins.py b/spaces/aodianyun/stable-diffusion-webui/extensions/deforum/scripts/deforum_helpers/src/adabins/unet_adaptive_bins.py
deleted file mode 100644
index 733927795146fe13563d07d20fbb461da596a181..0000000000000000000000000000000000000000
--- a/spaces/aodianyun/stable-diffusion-webui/extensions/deforum/scripts/deforum_helpers/src/adabins/unet_adaptive_bins.py
+++ /dev/null
@@ -1,154 +0,0 @@
-import torch
-import torch.nn as nn
-import torch.nn.functional as F
-import os
-from pathlib import Path
-
-from .miniViT import mViT
-
-
-class UpSampleBN(nn.Module):
- def __init__(self, skip_input, output_features):
- super(UpSampleBN, self).__init__()
-
- self._net = nn.Sequential(nn.Conv2d(skip_input, output_features, kernel_size=3, stride=1, padding=1),
- nn.BatchNorm2d(output_features),
- nn.LeakyReLU(),
- nn.Conv2d(output_features, output_features, kernel_size=3, stride=1, padding=1),
- nn.BatchNorm2d(output_features),
- nn.LeakyReLU())
-
- def forward(self, x, concat_with):
- up_x = F.interpolate(x, size=[concat_with.size(2), concat_with.size(3)], mode='bilinear', align_corners=True)
- f = torch.cat([up_x, concat_with], dim=1)
- return self._net(f)
-
-
-class DecoderBN(nn.Module):
- def __init__(self, num_features=2048, num_classes=1, bottleneck_features=2048):
- super(DecoderBN, self).__init__()
- features = int(num_features)
-
- self.conv2 = nn.Conv2d(bottleneck_features, features, kernel_size=1, stride=1, padding=1)
-
- self.up1 = UpSampleBN(skip_input=features // 1 + 112 + 64, output_features=features // 2)
- self.up2 = UpSampleBN(skip_input=features // 2 + 40 + 24, output_features=features // 4)
- self.up3 = UpSampleBN(skip_input=features // 4 + 24 + 16, output_features=features // 8)
- self.up4 = UpSampleBN(skip_input=features // 8 + 16 + 8, output_features=features // 16)
-
- # self.up5 = UpSample(skip_input=features // 16 + 3, output_features=features//16)
- self.conv3 = nn.Conv2d(features // 16, num_classes, kernel_size=3, stride=1, padding=1)
- # self.act_out = nn.Softmax(dim=1) if output_activation == 'softmax' else nn.Identity()
-
- def forward(self, features):
- x_block0, x_block1, x_block2, x_block3, x_block4 = features[4], features[5], features[6], features[8], features[
- 11]
-
- x_d0 = self.conv2(x_block4)
-
- x_d1 = self.up1(x_d0, x_block3)
- x_d2 = self.up2(x_d1, x_block2)
- x_d3 = self.up3(x_d2, x_block1)
- x_d4 = self.up4(x_d3, x_block0)
- # x_d5 = self.up5(x_d4, features[0])
- out = self.conv3(x_d4)
- # out = self.act_out(out)
- # if with_features:
- # return out, features[-1]
- # elif with_intermediate:
- # return out, [x_block0, x_block1, x_block2, x_block3, x_block4, x_d1, x_d2, x_d3, x_d4]
- return out
-
-
-class Encoder(nn.Module):
- def __init__(self, backend):
- super(Encoder, self).__init__()
- self.original_model = backend
-
- def forward(self, x):
- features = [x]
- for k, v in self.original_model._modules.items():
- if (k == 'blocks'):
- for ki, vi in v._modules.items():
- features.append(vi(features[-1]))
- else:
- features.append(v(features[-1]))
- return features
-
-
-class UnetAdaptiveBins(nn.Module):
- def __init__(self, backend, n_bins=100, min_val=0.1, max_val=10, norm='linear'):
- super(UnetAdaptiveBins, self).__init__()
- self.num_classes = n_bins
- self.min_val = min_val
- self.max_val = max_val
- self.encoder = Encoder(backend)
- self.adaptive_bins_layer = mViT(128, n_query_channels=128, patch_size=16,
- dim_out=n_bins,
- embedding_dim=128, norm=norm)
-
- self.decoder = DecoderBN(num_classes=128)
- self.conv_out = nn.Sequential(nn.Conv2d(128, n_bins, kernel_size=1, stride=1, padding=0),
- nn.Softmax(dim=1))
-
- def forward(self, x, **kwargs):
- unet_out = self.decoder(self.encoder(x), **kwargs)
- bin_widths_normed, range_attention_maps = self.adaptive_bins_layer(unet_out)
- out = self.conv_out(range_attention_maps)
-
- # Post process
- # n, c, h, w = out.shape
- # hist = torch.sum(out.view(n, c, h * w), dim=2) / (h * w) # not used for training
-
- bin_widths = (self.max_val - self.min_val) * bin_widths_normed # .shape = N, dim_out
- bin_widths = nn.functional.pad(bin_widths, (1, 0), mode='constant', value=self.min_val)
- bin_edges = torch.cumsum(bin_widths, dim=1)
-
- centers = 0.5 * (bin_edges[:, :-1] + bin_edges[:, 1:])
- n, dout = centers.size()
- centers = centers.view(n, dout, 1, 1)
-
- pred = torch.sum(out * centers, dim=1, keepdim=True)
-
- return bin_edges, pred
-
- def get_1x_lr_params(self): # lr/10 learning rate
- return self.encoder.parameters()
-
- def get_10x_lr_params(self): # lr learning rate
- modules = [self.decoder, self.adaptive_bins_layer, self.conv_out]
- for m in modules:
- yield from m.parameters()
-
- @classmethod
- def build(cls, n_bins, **kwargs):
- basemodel_name = 'tf_efficientnet_b5_ap'
-
- print('Loading base model ()...'.format(basemodel_name), end='')
- predicted_torch_model_cache_path = str(Path.home()) + '\\.cache\\torch\\hub\\rwightman_gen-efficientnet-pytorch_master'
- predicted_gep_cache_testilfe = Path(predicted_torch_model_cache_path + '\\hubconf.py')
- #print(f"predicted_gep_cache_testilfe: {predicted_gep_cache_testilfe}")
- # try to fetch the models from cache, and only if it can't be find, download from the internet (to enable offline usage)
- if os.path.isfile(predicted_gep_cache_testilfe):
- basemodel = torch.hub.load(predicted_torch_model_cache_path, basemodel_name, pretrained=True, source = 'local')
- else:
- basemodel = torch.hub.load('rwightman/gen-efficientnet-pytorch', basemodel_name, pretrained=True)
- print('Done.')
-
- # Remove last layer
- print('Removing last two layers (global_pool & classifier).')
- basemodel.global_pool = nn.Identity()
- basemodel.classifier = nn.Identity()
-
- # Building Encoder-Decoder model
- print('Building Encoder-Decoder model..', end='')
- m = cls(basemodel, n_bins=n_bins, **kwargs)
- print('Done.')
- return m
-
-
-if __name__ == '__main__':
- model = UnetAdaptiveBins.build(100)
- x = torch.rand(2, 3, 480, 640)
- bins, pred = model(x)
- print(bins.shape, pred.shape)
diff --git a/spaces/arslan-ahmed/talk-to-your-docs/whatsapp_chat_custom.py b/spaces/arslan-ahmed/talk-to-your-docs/whatsapp_chat_custom.py
deleted file mode 100644
index 39d5762f8e57399f75fffa609c6b7c07bfaeb669..0000000000000000000000000000000000000000
--- a/spaces/arslan-ahmed/talk-to-your-docs/whatsapp_chat_custom.py
+++ /dev/null
@@ -1,49 +0,0 @@
-# created custom class for WhatsAppChatLoader - because original langchain one isnt working
-
-import re
-from pathlib import Path
-from typing import List
-
-from langchain.docstore.document import Document
-from langchain.document_loaders.base import BaseLoader
-
-
-def concatenate_rows(date: str, sender: str, text: str) -> str:
- """Combine message information in a readable format ready to be used."""
- return f"{sender} on {date}: {text}\n\n"
-
-# def concatenate_rows(date: str, sender: str, text: str) -> str:
-# """Combine message information in a readable format ready to be used."""
-# return f"{text}\n"
-
-class WhatsAppChatLoader(BaseLoader):
- """Load `WhatsApp` messages text file."""
-
- def __init__(self, path: str):
- """Initialize with path."""
- self.file_path = path
-
- def load(self) -> List[Document]:
- """Load documents."""
- p = Path(self.file_path)
- text_content = ""
-
- ignore_lines = ["This message was deleted", ""]
- #########################################################################################
- # original code from langchain replaced with this code
- #########################################################################################
- # use https://whatstk.streamlit.app/ to get CSV
- import pandas as pd
- df = pd.read_csv(p)[['date', 'username', 'message']]
-
- for i,row in df.iterrows():
- date = row['date']
- sender = row['username']
- text = row['message']
-
- if not any(x in text for x in ignore_lines):
- text_content += concatenate_rows(date, sender, text)
-
- metadata = {"source": str(p)}
-
- return [Document(page_content=text_content.strip(), metadata=metadata)]
\ No newline at end of file
diff --git a/spaces/artificialguybr/video-dubbing/TTS/docs/source/tts_datasets.md b/spaces/artificialguybr/video-dubbing/TTS/docs/source/tts_datasets.md
deleted file mode 100644
index 11da1b7688d07dadfdb3dfab33deb4bcdf3f861a..0000000000000000000000000000000000000000
--- a/spaces/artificialguybr/video-dubbing/TTS/docs/source/tts_datasets.md
+++ /dev/null
@@ -1,17 +0,0 @@
-# TTS Datasets
-
-Some of the known public datasets that we successfully applied 🐸TTS:
-
-- [English - LJ Speech](https://keithito.com/LJ-Speech-Dataset/)
-- [English - Nancy](http://www.cstr.ed.ac.uk/projects/blizzard/2011/lessac_blizzard2011/)
-- [English - TWEB](https://www.kaggle.com/bryanpark/the-world-english-bible-speech-dataset)
-- [English - LibriTTS](https://openslr.org/60/)
-- [English - VCTK](https://datashare.ed.ac.uk/handle/10283/2950)
-- [Multilingual - M-AI-Labs](http://www.caito.de/2019/01/the-m-ailabs-speech-dataset/)
-- [Spanish](https://drive.google.com/file/d/1Sm_zyBo67XHkiFhcRSQ4YaHPYM0slO_e/view?usp=sharing) - thx! @carlfm01
-- [German - Thorsten OGVD](https://github.com/thorstenMueller/deep-learning-german-tts)
-- [Japanese - Kokoro](https://www.kaggle.com/kaiida/kokoro-speech-dataset-v11-small/version/1)
-- [Chinese](https://www.data-baker.com/data/index/source/)
-- [Ukrainian - LADA](https://github.com/egorsmkv/ukrainian-tts-datasets/tree/main/lada)
-
-Let us know if you use 🐸TTS on a different dataset.
diff --git a/spaces/asimokby/cv-parser-huggingface/ResumeReader.py b/spaces/asimokby/cv-parser-huggingface/ResumeReader.py
deleted file mode 100644
index e122a299de5d0a30b6c5b44c166514e19ac089fa..0000000000000000000000000000000000000000
--- a/spaces/asimokby/cv-parser-huggingface/ResumeReader.py
+++ /dev/null
@@ -1,99 +0,0 @@
-import re
-import os
-import logging
-import pdfplumber
-
-class ResumeReader:
-
- def convert_docx_to_txt(self, docx_file,docx_parser):
- """
- A utility function to convert a Microsoft docx files to raw text.
-
- This code is largely borrowed from existing solutions, and does not match the style of the rest of this repo.
- :param docx_file: docx file with gets uploaded by the user
- :type docx_file: InMemoryUploadedFile
- :return: The text contents of the docx file
- :rtype: str
- """
-
- # doc = docx.Document(docx_file)
- # allText = []
- # for docpara in doc.paragraphs:
- # allText.append(docpara.text)
- # text = ' '.join(allText)
- text = ""
- try:
- clean_text = re.sub(r'\n+', '\n', text)
- clean_text = clean_text.replace("\r", "\n").replace("\t", " ") # Normalize text blob
- resume_lines = clean_text.splitlines() # Split text blob into individual lines
- resume_lines = [re.sub('\s+', ' ', line.strip()) for line in resume_lines if
- line.strip()] # Remove empty strings and whitespaces
- return resume_lines, text
- except Exception as e:
- logging.error('Error in docx file:: ' + str(e))
- return [], " "
-
- def convert_pdf_to_txt(self, pdf_file):
- """
- A utility function to convert a machine-readable PDF to raw text.
-
- This code is largely borrowed from existing solutions, and does not match the style of the rest of this repo.
- :param input_pdf_path: Path to the .pdf file which should be converted
- :type input_pdf_path: str
- :return: The text contents of the pdf
- :rtype: str
- """
-
- pdf = pdfplumber.open(pdf_file)
- raw_text= ""
-
- for page in pdf.pages:
- raw_text += page.extract_text() + "\n"
-
- pdf.close()
-
- try:
- full_string = re.sub(r'\n+', '\n', raw_text)
- full_string = full_string.replace("\r", "\n")
- full_string = full_string.replace("\t", " ")
-
- # Remove awkward LaTeX bullet characters
- full_string = re.sub(r"\uf0b7", " ", full_string)
- full_string = re.sub(r"\(cid:\d{0,3}\)", " ", full_string)
- full_string = re.sub(r'• ', " ", full_string)
-
- # Split text blob into individual lines
- resume_lines = full_string.splitlines(True)
-
- # Remove empty strings and whitespaces
- resume_lines = [re.sub('\s+', ' ', line.strip()) for line in resume_lines if line.strip()]
-
- return resume_lines, raw_text
- except Exception as e:
- logging.error('Error in docx file:: ' + str(e))
- return [], " "
-
- def read_file(self, file,docx_parser = "tika"):
- """
- file : Give path of resume file
- docx_parser : Enter docx2txt or tika, by default is tika
- """
- print("Reading the Resume...")
- # file = "/content/Asst Manager Trust Administration.docx"
- file = os.path.join(file)
- if file.endswith('docx') or file.endswith('doc'):
- # if file.endswith('doc') and docx_parser == "docx2txt":
- # docx_parser = "tika"
- # logging.error("doc format not supported by the docx2txt changing back to tika")
- resume_lines, raw_text = self.convert_docx_to_txt(file,docx_parser)
- elif file.endswith('pdf'):
- resume_lines, raw_text = self.convert_pdf_to_txt(file)
- elif file.endswith('txt'):
- with open(file, 'r', encoding='utf-8') as f:
- resume_lines = f.readlines()
-
- else:
- resume_lines = None
-
-
- return resume_lines
\ No newline at end of file
diff --git a/spaces/awacke1/AutoMLUsingStreamlit-Plotly/Prompts.md b/spaces/awacke1/AutoMLUsingStreamlit-Plotly/Prompts.md
deleted file mode 100644
index ec0817592c1a131230048d4f6e9994f6d2c9863d..0000000000000000000000000000000000000000
--- a/spaces/awacke1/AutoMLUsingStreamlit-Plotly/Prompts.md
+++ /dev/null
@@ -1,115 +0,0 @@
-
-
-write an automl streamlit python tensorflow keras application with mechanics that takes a CSV dataset, performs pandas profiling and generates the code to do classification
-Here is an example of an Automated Machine Learning (AutoML) Streamlit application that takes a CSV dataset, performs pandas profiling, and generates the code to do classification using TensorFlow and Keras:
-
-python
-Copy code
-import streamlit as st
-import pandas as pd
-import pandas_profiling
-from tensorflow.keras.models import Sequential
-from tensorflow.keras.layers import Dense
-from sklearn.model_selection import train_test_split
-
-st.set_page_config(page_title="AutoML Streamlit App", page_icon=":robot:", layout="wide")
-
-st.title("AutoML Streamlit App")
-
-# Upload a CSV dataset
-uploaded_file = st.file_uploader("Upload your dataset", type=["csv"])
-if uploaded_file is not None:
- # Load the dataset and display the first 5 rows
- df = pd.read_csv(uploaded_file)
- st.dataframe(df.head())
-
- # Perform pandas profiling and display the report
- profile = pandas_profiling.ProfileReport(df)
- st.write("Pandas Profiling Report:", profile)
-
- # Split the data into features and target variables
- X = df.drop(df.columns[-1], axis=1)
- y = df[df.columns[-1]]
-
- # Split the data into training and testing sets
- X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2)
-
- # Define a sequential model with TensorFlow and Keras
- model = Sequential()
- model.add(Dense(16, input_dim=X_train.shape[1], activation="relu"))
- model.add(Dense(1, activation="sigmoid"))
-
- # Compile the model
- model.compile(loss="binary_crossentropy", optimizer="adam", metrics=["accuracy"])
-
- # Train the model
- model.fit(X_train, y_train, epochs=10, batch_size=32)
-
- # Evaluate the model on the test set
- test_loss, test_accuracy = model.evaluate(X_test, y_test)
- st.write("Test Accuracy:", test_accuracy)
-
- # Generate the code to do classification
- st.write("Generated Code:")
- st.code("""
-import pandas as pd
-import tensorflow as tf
-from tensorflow.keras.models import Sequential
-from tensorflow.keras.layers import Dense
-
-# Load the dataset
-df = pd.read_csv("dataset.csv")
-
-# Split the data into features and target variables
-X = df.drop(df.columns[-1], axis=1)
-y = df[df.columns[-1]]
-
-# Define a sequential model with TensorFlow and Keras
-model = Sequential()
-model.add(Dense(16, input_dim=X.shape[1], activation="relu"))
-model.add(Dense(1, activation="sigmoid"))
-
-# Compile the model
-model.compile(loss="binary_crossentropy", optimizer="
-
-
-
-Aaron Wacker
-continue writing the program above but simplify it and automatically generate a plotly visualization using treemap and sunburst plot from the dataset dynamically by interpreting the data types inside the CSV file
-Here is a simplified version of the AutoML Streamlit application that automatically generates a Plotly visualization using treemap and sunburst plots from the CSV dataset dynamically by interpreting the data types inside the file:
-
-python
-Copy code
-import streamlit as st
-import pandas as pd
-import plotly.express as px
-
-st.set_page_config(page_title="AutoML Streamlit App", page_icon=":robot:", layout="wide")
-
-st.title("AutoML Streamlit App")
-
-# Upload a CSV dataset
-uploaded_file = st.file_uploader("Upload your dataset", type=["csv"])
-if uploaded_file is not None:
- # Load the dataset and display the first 5 rows
- df = pd.read_csv(uploaded_file)
- st.dataframe(df.head())
-
- # Generate a treemap or sunburst plot based on data types
- numerical_cols = df.select_dtypes(include=["float", "int"]).columns
- categorical_cols = df.select_dtypes(include=["object"]).columns
-
- if len(numerical_cols) >= 2:
- fig = px.scatter_matrix(df, dimensions=numerical_cols)
- st.plotly_chart(fig)
- elif len(categorical_cols) >= 2:
- fig = px.treemap(df, path=categorical_cols)
- st.plotly_chart(fig)
- else:
- fig = px.sunburst(df, path=categorical_cols + numerical_cols)
- st.plotly_chart(fig)
-In this simplified version, we use Plotly Express to generate a scatter matrix plot for numerical data, a treemap plot for categorical data, and a sunburst plot for both numerical and categorical data. The type of plot generated is based on the number of numerical and categorical columns present in the dataset.
-
-
-
-Regenerate response
diff --git a/spaces/awacke1/AutoStableDiffusionTxt2ImgImg2Img/README.md b/spaces/awacke1/AutoStableDiffusionTxt2ImgImg2Img/README.md
deleted file mode 100644
index d09d8ce162e139ce06f130f29b73cd0221407ed6..0000000000000000000000000000000000000000
--- a/spaces/awacke1/AutoStableDiffusionTxt2ImgImg2Img/README.md
+++ /dev/null
@@ -1,20 +0,0 @@
----
-title: Stable Diffusion Web UI Docker
-emoji: 🐳
-colorFrom: blue
-colorTo: blue
-sdk: docker
-sdk_version: 3.9
-app_file: oh-no.py
-pinned: false
-duplicated_from: camenduru/webui-docker
----
-
-## Stable Diffusion Web UI
-https://github.com/AUTOMATIC1111/stable-diffusion-webui
-
-## Documentation
-https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki
-
-## Models License
-https://huggingface.co/spaces/CompVis/stable-diffusion-license
\ No newline at end of file
diff --git a/spaces/awacke1/Docker.VSCode.Integration.HF/README.md b/spaces/awacke1/Docker.VSCode.Integration.HF/README.md
deleted file mode 100644
index 081c2aca5247737651bb3972e5352ab8cc8aea90..0000000000000000000000000000000000000000
--- a/spaces/awacke1/Docker.VSCode.Integration.HF/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Visual Studio Code
-emoji: 💻🐳
-colorFrom: red
-colorTo: blue
-sdk: docker
-pinned: false
-tags:
-- vscode
-duplicated_from: DockerTemplates/vscode
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
\ No newline at end of file
diff --git a/spaces/awacke1/HTML5-Tower-Building-3D-Game/v1-index.html b/spaces/awacke1/HTML5-Tower-Building-3D-Game/v1-index.html
deleted file mode 100644
index bbf3eac2ba2fd26e3aa6c1e182b287a3e242e878..0000000000000000000000000000000000000000
--- a/spaces/awacke1/HTML5-Tower-Building-3D-Game/v1-index.html
+++ /dev/null
@@ -1,101 +0,0 @@
-
-
-
- Tower Building Game
-
-
-
-
-
-
-
-
diff --git a/spaces/awacke1/VizLib-Keras-n-Plotly/README.md b/spaces/awacke1/VizLib-Keras-n-Plotly/README.md
deleted file mode 100644
index 6d0928d18a928cf4ec2a6a4617eb388de43a9fcb..0000000000000000000000000000000000000000
--- a/spaces/awacke1/VizLib-Keras-n-Plotly/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: VizLib Keras N Plotly
-emoji: 💻
-colorFrom: red
-colorTo: indigo
-sdk: streamlit
-sdk_version: 1.17.0
-app_file: app.py
-pinned: false
-license: mit
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/awacke1/WebAssemblyStreamlitLite-stlite/index.html b/spaces/awacke1/WebAssemblyStreamlitLite-stlite/index.html
deleted file mode 100644
index 44bd1cc51118aa69255d585846dc448f5f61e97a..0000000000000000000000000000000000000000
--- a/spaces/awacke1/WebAssemblyStreamlitLite-stlite/index.html
+++ /dev/null
@@ -1,33 +0,0 @@
-
-
-
-
-
- WebAssemblyStreamlitLite-stlite in HTML5
-
-
-
-
-
-
DarkEbootFixerV55rar Gwen!'s Board. - Explore Gwen!'s board. DarkEbootFixerV55rar you've just visited the back page. It's dark so you can't read the URL, but for sure you can look over your shoulder. The DarkEbootFixerV55rar team is proud to announce the new version of DMP Boot Fix Download. DarkEbootFixerV55rar, in case you don't know, is a tool that repairs issues related to DMP media. DarkEbootFixerV55rar have a log for each version of DMP media, so that it is possible to compare the two versions. That way, if there is a difference, you can safely download and install the new version of DMP Boot Fix. The new version, DMP Boot Fix Download, fixed a few bugs related to DMP media. The version is 0.9.2 and is fully compatible with most DMP media formats. The developers have also added a compatibility checker, so you can easily determine whether your media is compatible with the latest version of the software. Download DarkEbootFixerV55rar, it's free and safe and includes a one year warranty. The DarkEbootFixerV55rar team has also released a free release of DMP Remote Link, version 0.92. https://coub.com/stories/7649178-darkebootfixerv55rar-free-release-compatibility-check. http://codex.downloaddarkbootfixer.
charn6 0d958a46dc https://coub.com/stories/5260461-darkebootfixerv55rar-switch. yoshitano 24 1. 1.0 You have to use it 2.) The same as you may right now. 2. I accidentally get the iso files of the v55 RAM replacement which is.
899543212b
-
-
\ No newline at end of file
diff --git a/spaces/blmdsydm/faster-whisper-webui/app-local.py b/spaces/blmdsydm/faster-whisper-webui/app-local.py
deleted file mode 100644
index c7717d096ca5f95177f0dba03cd62ca729bae9f3..0000000000000000000000000000000000000000
--- a/spaces/blmdsydm/faster-whisper-webui/app-local.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# Run the app with no audio file restrictions
-from app import create_ui
-from src.config import ApplicationConfig
-
-create_ui(ApplicationConfig.create_default(input_audio_max_duration=-1))
\ No newline at end of file
diff --git a/spaces/brjathu/HMR2.0/vendor/detectron2/detectron2/utils/memory.py b/spaces/brjathu/HMR2.0/vendor/detectron2/detectron2/utils/memory.py
deleted file mode 100644
index bd494780b9dbbd1571688cd270bb9b53d113c13e..0000000000000000000000000000000000000000
--- a/spaces/brjathu/HMR2.0/vendor/detectron2/detectron2/utils/memory.py
+++ /dev/null
@@ -1,84 +0,0 @@
-# Copyright (c) Facebook, Inc. and its affiliates.
-
-import logging
-from contextlib import contextmanager
-from functools import wraps
-import torch
-
-__all__ = ["retry_if_cuda_oom"]
-
-
-@contextmanager
-def _ignore_torch_cuda_oom():
- """
- A context which ignores CUDA OOM exception from pytorch.
- """
- try:
- yield
- except RuntimeError as e:
- # NOTE: the string may change?
- if "CUDA out of memory. " in str(e):
- pass
- else:
- raise
-
-
-def retry_if_cuda_oom(func):
- """
- Makes a function retry itself after encountering
- pytorch's CUDA OOM error.
- It will first retry after calling `torch.cuda.empty_cache()`.
-
- If that still fails, it will then retry by trying to convert inputs to CPUs.
- In this case, it expects the function to dispatch to CPU implementation.
- The return values may become CPU tensors as well and it's user's
- responsibility to convert it back to CUDA tensor if needed.
-
- Args:
- func: a stateless callable that takes tensor-like objects as arguments
-
- Returns:
- a callable which retries `func` if OOM is encountered.
-
- Examples:
- ::
- output = retry_if_cuda_oom(some_torch_function)(input1, input2)
- # output may be on CPU even if inputs are on GPU
-
- Note:
- 1. When converting inputs to CPU, it will only look at each argument and check
- if it has `.device` and `.to` for conversion. Nested structures of tensors
- are not supported.
-
- 2. Since the function might be called more than once, it has to be
- stateless.
- """
-
- def maybe_to_cpu(x):
- try:
- like_gpu_tensor = x.device.type == "cuda" and hasattr(x, "to")
- except AttributeError:
- like_gpu_tensor = False
- if like_gpu_tensor:
- return x.to(device="cpu")
- else:
- return x
-
- @wraps(func)
- def wrapped(*args, **kwargs):
- with _ignore_torch_cuda_oom():
- return func(*args, **kwargs)
-
- # Clear cache and retry
- torch.cuda.empty_cache()
- with _ignore_torch_cuda_oom():
- return func(*args, **kwargs)
-
- # Try on CPU. This slows down the code significantly, therefore print a notice.
- logger = logging.getLogger(__name__)
- logger.info("Attempting to copy inputs of {} to CPU due to CUDA OOM".format(str(func)))
- new_args = (maybe_to_cpu(x) for x in args)
- new_kwargs = {k: maybe_to_cpu(v) for k, v in kwargs.items()}
- return func(*new_args, **new_kwargs)
-
- return wrapped
diff --git a/spaces/brjathu/HMR2.0/vendor/detectron2/projects/DeepLab/deeplab/resnet.py b/spaces/brjathu/HMR2.0/vendor/detectron2/projects/DeepLab/deeplab/resnet.py
deleted file mode 100644
index 2cc277b24630a9425f4c37e1abc3352b49e1a031..0000000000000000000000000000000000000000
--- a/spaces/brjathu/HMR2.0/vendor/detectron2/projects/DeepLab/deeplab/resnet.py
+++ /dev/null
@@ -1,158 +0,0 @@
-# Copyright (c) Facebook, Inc. and its affiliates.
-import fvcore.nn.weight_init as weight_init
-import torch.nn.functional as F
-
-from detectron2.layers import CNNBlockBase, Conv2d, get_norm
-from detectron2.modeling import BACKBONE_REGISTRY
-from detectron2.modeling.backbone.resnet import (
- BasicStem,
- BottleneckBlock,
- DeformBottleneckBlock,
- ResNet,
-)
-
-
-class DeepLabStem(CNNBlockBase):
- """
- The DeepLab ResNet stem (layers before the first residual block).
- """
-
- def __init__(self, in_channels=3, out_channels=128, norm="BN"):
- """
- Args:
- norm (str or callable): norm after the first conv layer.
- See :func:`layers.get_norm` for supported format.
- """
- super().__init__(in_channels, out_channels, 4)
- self.in_channels = in_channels
- self.conv1 = Conv2d(
- in_channels,
- out_channels // 2,
- kernel_size=3,
- stride=2,
- padding=1,
- bias=False,
- norm=get_norm(norm, out_channels // 2),
- )
- self.conv2 = Conv2d(
- out_channels // 2,
- out_channels // 2,
- kernel_size=3,
- stride=1,
- padding=1,
- bias=False,
- norm=get_norm(norm, out_channels // 2),
- )
- self.conv3 = Conv2d(
- out_channels // 2,
- out_channels,
- kernel_size=3,
- stride=1,
- padding=1,
- bias=False,
- norm=get_norm(norm, out_channels),
- )
- weight_init.c2_msra_fill(self.conv1)
- weight_init.c2_msra_fill(self.conv2)
- weight_init.c2_msra_fill(self.conv3)
-
- def forward(self, x):
- x = self.conv1(x)
- x = F.relu_(x)
- x = self.conv2(x)
- x = F.relu_(x)
- x = self.conv3(x)
- x = F.relu_(x)
- x = F.max_pool2d(x, kernel_size=3, stride=2, padding=1)
- return x
-
-
-@BACKBONE_REGISTRY.register()
-def build_resnet_deeplab_backbone(cfg, input_shape):
- """
- Create a ResNet instance from config.
- Returns:
- ResNet: a :class:`ResNet` instance.
- """
- # need registration of new blocks/stems?
- norm = cfg.MODEL.RESNETS.NORM
- if cfg.MODEL.RESNETS.STEM_TYPE == "basic":
- stem = BasicStem(
- in_channels=input_shape.channels,
- out_channels=cfg.MODEL.RESNETS.STEM_OUT_CHANNELS,
- norm=norm,
- )
- elif cfg.MODEL.RESNETS.STEM_TYPE == "deeplab":
- stem = DeepLabStem(
- in_channels=input_shape.channels,
- out_channels=cfg.MODEL.RESNETS.STEM_OUT_CHANNELS,
- norm=norm,
- )
- else:
- raise ValueError("Unknown stem type: {}".format(cfg.MODEL.RESNETS.STEM_TYPE))
-
- # fmt: off
- freeze_at = cfg.MODEL.BACKBONE.FREEZE_AT
- out_features = cfg.MODEL.RESNETS.OUT_FEATURES
- depth = cfg.MODEL.RESNETS.DEPTH
- num_groups = cfg.MODEL.RESNETS.NUM_GROUPS
- width_per_group = cfg.MODEL.RESNETS.WIDTH_PER_GROUP
- bottleneck_channels = num_groups * width_per_group
- in_channels = cfg.MODEL.RESNETS.STEM_OUT_CHANNELS
- out_channels = cfg.MODEL.RESNETS.RES2_OUT_CHANNELS
- stride_in_1x1 = cfg.MODEL.RESNETS.STRIDE_IN_1X1
- res4_dilation = cfg.MODEL.RESNETS.RES4_DILATION
- res5_dilation = cfg.MODEL.RESNETS.RES5_DILATION
- deform_on_per_stage = cfg.MODEL.RESNETS.DEFORM_ON_PER_STAGE
- deform_modulated = cfg.MODEL.RESNETS.DEFORM_MODULATED
- deform_num_groups = cfg.MODEL.RESNETS.DEFORM_NUM_GROUPS
- res5_multi_grid = cfg.MODEL.RESNETS.RES5_MULTI_GRID
- # fmt: on
- assert res4_dilation in {1, 2}, "res4_dilation cannot be {}.".format(res4_dilation)
- assert res5_dilation in {1, 2, 4}, "res5_dilation cannot be {}.".format(res5_dilation)
- if res4_dilation == 2:
- # Always dilate res5 if res4 is dilated.
- assert res5_dilation == 4
-
- num_blocks_per_stage = {50: [3, 4, 6, 3], 101: [3, 4, 23, 3], 152: [3, 8, 36, 3]}[depth]
-
- stages = []
-
- # Avoid creating variables without gradients
- # It consumes extra memory and may cause allreduce to fail
- out_stage_idx = [{"res2": 2, "res3": 3, "res4": 4, "res5": 5}[f] for f in out_features]
- max_stage_idx = max(out_stage_idx)
- for idx, stage_idx in enumerate(range(2, max_stage_idx + 1)):
- if stage_idx == 4:
- dilation = res4_dilation
- elif stage_idx == 5:
- dilation = res5_dilation
- else:
- dilation = 1
- first_stride = 1 if idx == 0 or dilation > 1 else 2
- stage_kargs = {
- "num_blocks": num_blocks_per_stage[idx],
- "stride_per_block": [first_stride] + [1] * (num_blocks_per_stage[idx] - 1),
- "in_channels": in_channels,
- "out_channels": out_channels,
- "norm": norm,
- }
- stage_kargs["bottleneck_channels"] = bottleneck_channels
- stage_kargs["stride_in_1x1"] = stride_in_1x1
- stage_kargs["dilation"] = dilation
- stage_kargs["num_groups"] = num_groups
- if deform_on_per_stage[idx]:
- stage_kargs["block_class"] = DeformBottleneckBlock
- stage_kargs["deform_modulated"] = deform_modulated
- stage_kargs["deform_num_groups"] = deform_num_groups
- else:
- stage_kargs["block_class"] = BottleneckBlock
- if stage_idx == 5:
- stage_kargs.pop("dilation")
- stage_kargs["dilation_per_block"] = [dilation * mg for mg in res5_multi_grid]
- blocks = ResNet.make_stage(**stage_kargs)
- in_channels = out_channels
- out_channels *= 2
- bottleneck_channels *= 2
- stages.append(blocks)
- return ResNet(stem, stages, out_features=out_features).freeze(freeze_at)
diff --git a/spaces/bulentsofttech/gradio_s1000_veri_toplama_modeli/yolov5/utils/plots.py b/spaces/bulentsofttech/gradio_s1000_veri_toplama_modeli/yolov5/utils/plots.py
deleted file mode 100644
index 1bbb9c09c33afe83c90d6ea96511ae64c8d9bec9..0000000000000000000000000000000000000000
--- a/spaces/bulentsofttech/gradio_s1000_veri_toplama_modeli/yolov5/utils/plots.py
+++ /dev/null
@@ -1,489 +0,0 @@
-# YOLOv5 🚀 by Ultralytics, GPL-3.0 license
-"""
-Plotting utils
-"""
-
-import math
-import os
-from copy import copy
-from pathlib import Path
-from urllib.error import URLError
-
-import cv2
-import matplotlib
-import matplotlib.pyplot as plt
-import numpy as np
-import pandas as pd
-import seaborn as sn
-import torch
-from PIL import Image, ImageDraw, ImageFont
-
-from utils.general import (CONFIG_DIR, FONT, LOGGER, Timeout, check_font, check_requirements, clip_coords,
- increment_path, is_ascii, threaded, try_except, xywh2xyxy, xyxy2xywh)
-from utils.metrics import fitness
-
-# Settings
-RANK = int(os.getenv('RANK', -1))
-matplotlib.rc('font', **{'size': 11})
-matplotlib.use('Agg') # for writing to files only
-
-
-class Colors:
- # Ultralytics color palette https://ultralytics.com/
- def __init__(self):
- # hex = matplotlib.colors.TABLEAU_COLORS.values()
- hexs = ('FF3838', 'FF9D97', 'FF701F', 'FFB21D', 'CFD231', '48F90A', '92CC17', '3DDB86', '1A9334', '00D4BB',
- '2C99A8', '00C2FF', '344593', '6473FF', '0018EC', '8438FF', '520085', 'CB38FF', 'FF95C8', 'FF37C7')
- self.palette = [self.hex2rgb(f'#{c}') for c in hexs]
- self.n = len(self.palette)
-
- def __call__(self, i, bgr=False):
- c = self.palette[int(i) % self.n]
- return (c[2], c[1], c[0]) if bgr else c
-
- @staticmethod
- def hex2rgb(h): # rgb order (PIL)
- return tuple(int(h[1 + i:1 + i + 2], 16) for i in (0, 2, 4))
-
-
-colors = Colors() # create instance for 'from utils.plots import colors'
-
-
-def check_pil_font(font=FONT, size=10):
- # Return a PIL TrueType Font, downloading to CONFIG_DIR if necessary
- font = Path(font)
- font = font if font.exists() else (CONFIG_DIR / font.name)
- try:
- return ImageFont.truetype(str(font) if font.exists() else font.name, size)
- except Exception: # download if missing
- try:
- check_font(font)
- return ImageFont.truetype(str(font), size)
- except TypeError:
- check_requirements('Pillow>=8.4.0') # known issue https://github.com/ultralytics/yolov5/issues/5374
- except URLError: # not online
- return ImageFont.load_default()
-
-
-class Annotator:
- # YOLOv5 Annotator for train/val mosaics and jpgs and detect/hub inference annotations
- def __init__(self, im, line_width=None, font_size=None, font='Arial.ttf', pil=False, example='abc'):
- assert im.data.contiguous, 'Image not contiguous. Apply np.ascontiguousarray(im) to Annotator() input images.'
- non_ascii = not is_ascii(example) # non-latin labels, i.e. asian, arabic, cyrillic
- self.pil = pil or non_ascii
- if self.pil: # use PIL
- self.im = im if isinstance(im, Image.Image) else Image.fromarray(im)
- self.draw = ImageDraw.Draw(self.im)
- self.font = check_pil_font(font='Arial.Unicode.ttf' if non_ascii else font,
- size=font_size or max(round(sum(self.im.size) / 2 * 0.035), 12))
- else: # use cv2
- self.im = im
- self.lw = line_width or max(round(sum(im.shape) / 2 * 0.003), 2) # line width
-
- def box_label(self, box, label='', color=(128, 128, 128), txt_color=(255, 255, 255)):
- # Add one xyxy box to image with label
- if self.pil or not is_ascii(label):
- self.draw.rectangle(box, width=self.lw, outline=color) # box
- if label:
- w, h = self.font.getsize(label) # text width, height
- outside = box[1] - h >= 0 # label fits outside box
- self.draw.rectangle(
- (box[0], box[1] - h if outside else box[1], box[0] + w + 1,
- box[1] + 1 if outside else box[1] + h + 1),
- fill=color,
- )
- # self.draw.text((box[0], box[1]), label, fill=txt_color, font=self.font, anchor='ls') # for PIL>8.0
- self.draw.text((box[0], box[1] - h if outside else box[1]), label, fill=txt_color, font=self.font)
- else: # cv2
- p1, p2 = (int(box[0]), int(box[1])), (int(box[2]), int(box[3]))
- cv2.rectangle(self.im, p1, p2, color, thickness=self.lw, lineType=cv2.LINE_AA)
- if label:
- tf = max(self.lw - 1, 1) # font thickness
- w, h = cv2.getTextSize(label, 0, fontScale=self.lw / 3, thickness=tf)[0] # text width, height
- outside = p1[1] - h >= 3
- p2 = p1[0] + w, p1[1] - h - 3 if outside else p1[1] + h + 3
- cv2.rectangle(self.im, p1, p2, color, -1, cv2.LINE_AA) # filled
- cv2.putText(self.im,
- label, (p1[0], p1[1] - 2 if outside else p1[1] + h + 2),
- 0,
- self.lw / 3,
- txt_color,
- thickness=tf,
- lineType=cv2.LINE_AA)
-
- def rectangle(self, xy, fill=None, outline=None, width=1):
- # Add rectangle to image (PIL-only)
- self.draw.rectangle(xy, fill, outline, width)
-
- def text(self, xy, text, txt_color=(255, 255, 255)):
- # Add text to image (PIL-only)
- w, h = self.font.getsize(text) # text width, height
- self.draw.text((xy[0], xy[1] - h + 1), text, fill=txt_color, font=self.font)
-
- def result(self):
- # Return annotated image as array
- return np.asarray(self.im)
-
-
-def feature_visualization(x, module_type, stage, n=32, save_dir=Path('runs/detect/exp')):
- """
- x: Features to be visualized
- module_type: Module type
- stage: Module stage within model
- n: Maximum number of feature maps to plot
- save_dir: Directory to save results
- """
- if 'Detect' not in module_type:
- batch, channels, height, width = x.shape # batch, channels, height, width
- if height > 1 and width > 1:
- f = save_dir / f"stage{stage}_{module_type.split('.')[-1]}_features.png" # filename
-
- blocks = torch.chunk(x[0].cpu(), channels, dim=0) # select batch index 0, block by channels
- n = min(n, channels) # number of plots
- fig, ax = plt.subplots(math.ceil(n / 8), 8, tight_layout=True) # 8 rows x n/8 cols
- ax = ax.ravel()
- plt.subplots_adjust(wspace=0.05, hspace=0.05)
- for i in range(n):
- ax[i].imshow(blocks[i].squeeze()) # cmap='gray'
- ax[i].axis('off')
-
- LOGGER.info(f'Saving {f}... ({n}/{channels})')
- plt.savefig(f, dpi=300, bbox_inches='tight')
- plt.close()
- np.save(str(f.with_suffix('.npy')), x[0].cpu().numpy()) # npy save
-
-
-def hist2d(x, y, n=100):
- # 2d histogram used in labels.png and evolve.png
- xedges, yedges = np.linspace(x.min(), x.max(), n), np.linspace(y.min(), y.max(), n)
- hist, xedges, yedges = np.histogram2d(x, y, (xedges, yedges))
- xidx = np.clip(np.digitize(x, xedges) - 1, 0, hist.shape[0] - 1)
- yidx = np.clip(np.digitize(y, yedges) - 1, 0, hist.shape[1] - 1)
- return np.log(hist[xidx, yidx])
-
-
-def butter_lowpass_filtfilt(data, cutoff=1500, fs=50000, order=5):
- from scipy.signal import butter, filtfilt
-
- # https://stackoverflow.com/questions/28536191/how-to-filter-smooth-with-scipy-numpy
- def butter_lowpass(cutoff, fs, order):
- nyq = 0.5 * fs
- normal_cutoff = cutoff / nyq
- return butter(order, normal_cutoff, btype='low', analog=False)
-
- b, a = butter_lowpass(cutoff, fs, order=order)
- return filtfilt(b, a, data) # forward-backward filter
-
-
-def output_to_target(output):
- # Convert model output to target format [batch_id, class_id, x, y, w, h, conf]
- targets = []
- for i, o in enumerate(output):
- for *box, conf, cls in o.cpu().numpy():
- targets.append([i, cls, *list(*xyxy2xywh(np.array(box)[None])), conf])
- return np.array(targets)
-
-
-@threaded
-def plot_images(images, targets, paths=None, fname='images.jpg', names=None, max_size=1920, max_subplots=16):
- # Plot image grid with labels
- if isinstance(images, torch.Tensor):
- images = images.cpu().float().numpy()
- if isinstance(targets, torch.Tensor):
- targets = targets.cpu().numpy()
- if np.max(images[0]) <= 1:
- images *= 255 # de-normalise (optional)
- bs, _, h, w = images.shape # batch size, _, height, width
- bs = min(bs, max_subplots) # limit plot images
- ns = np.ceil(bs ** 0.5) # number of subplots (square)
-
- # Build Image
- mosaic = np.full((int(ns * h), int(ns * w), 3), 255, dtype=np.uint8) # init
- for i, im in enumerate(images):
- if i == max_subplots: # if last batch has fewer images than we expect
- break
- x, y = int(w * (i // ns)), int(h * (i % ns)) # block origin
- im = im.transpose(1, 2, 0)
- mosaic[y:y + h, x:x + w, :] = im
-
- # Resize (optional)
- scale = max_size / ns / max(h, w)
- if scale < 1:
- h = math.ceil(scale * h)
- w = math.ceil(scale * w)
- mosaic = cv2.resize(mosaic, tuple(int(x * ns) for x in (w, h)))
-
- # Annotate
- fs = int((h + w) * ns * 0.01) # font size
- annotator = Annotator(mosaic, line_width=round(fs / 10), font_size=fs, pil=True, example=names)
- for i in range(i + 1):
- x, y = int(w * (i // ns)), int(h * (i % ns)) # block origin
- annotator.rectangle([x, y, x + w, y + h], None, (255, 255, 255), width=2) # borders
- if paths:
- annotator.text((x + 5, y + 5 + h), text=Path(paths[i]).name[:40], txt_color=(220, 220, 220)) # filenames
- if len(targets) > 0:
- ti = targets[targets[:, 0] == i] # image targets
- boxes = xywh2xyxy(ti[:, 2:6]).T
- classes = ti[:, 1].astype('int')
- labels = ti.shape[1] == 6 # labels if no conf column
- conf = None if labels else ti[:, 6] # check for confidence presence (label vs pred)
-
- if boxes.shape[1]:
- if boxes.max() <= 1.01: # if normalized with tolerance 0.01
- boxes[[0, 2]] *= w # scale to pixels
- boxes[[1, 3]] *= h
- elif scale < 1: # absolute coords need scale if image scales
- boxes *= scale
- boxes[[0, 2]] += x
- boxes[[1, 3]] += y
- for j, box in enumerate(boxes.T.tolist()):
- cls = classes[j]
- color = colors(cls)
- cls = names[cls] if names else cls
- if labels or conf[j] > 0.25: # 0.25 conf thresh
- label = f'{cls}' if labels else f'{cls} {conf[j]:.1f}'
- annotator.box_label(box, label, color=color)
- annotator.im.save(fname) # save
-
-
-def plot_lr_scheduler(optimizer, scheduler, epochs=300, save_dir=''):
- # Plot LR simulating training for full epochs
- optimizer, scheduler = copy(optimizer), copy(scheduler) # do not modify originals
- y = []
- for _ in range(epochs):
- scheduler.step()
- y.append(optimizer.param_groups[0]['lr'])
- plt.plot(y, '.-', label='LR')
- plt.xlabel('epoch')
- plt.ylabel('LR')
- plt.grid()
- plt.xlim(0, epochs)
- plt.ylim(0)
- plt.savefig(Path(save_dir) / 'LR.png', dpi=200)
- plt.close()
-
-
-def plot_val_txt(): # from utils.plots import *; plot_val()
- # Plot val.txt histograms
- x = np.loadtxt('val.txt', dtype=np.float32)
- box = xyxy2xywh(x[:, :4])
- cx, cy = box[:, 0], box[:, 1]
-
- fig, ax = plt.subplots(1, 1, figsize=(6, 6), tight_layout=True)
- ax.hist2d(cx, cy, bins=600, cmax=10, cmin=0)
- ax.set_aspect('equal')
- plt.savefig('hist2d.png', dpi=300)
-
- fig, ax = plt.subplots(1, 2, figsize=(12, 6), tight_layout=True)
- ax[0].hist(cx, bins=600)
- ax[1].hist(cy, bins=600)
- plt.savefig('hist1d.png', dpi=200)
-
-
-def plot_targets_txt(): # from utils.plots import *; plot_targets_txt()
- # Plot targets.txt histograms
- x = np.loadtxt('targets.txt', dtype=np.float32).T
- s = ['x targets', 'y targets', 'width targets', 'height targets']
- fig, ax = plt.subplots(2, 2, figsize=(8, 8), tight_layout=True)
- ax = ax.ravel()
- for i in range(4):
- ax[i].hist(x[i], bins=100, label=f'{x[i].mean():.3g} +/- {x[i].std():.3g}')
- ax[i].legend()
- ax[i].set_title(s[i])
- plt.savefig('targets.jpg', dpi=200)
-
-
-def plot_val_study(file='', dir='', x=None): # from utils.plots import *; plot_val_study()
- # Plot file=study.txt generated by val.py (or plot all study*.txt in dir)
- save_dir = Path(file).parent if file else Path(dir)
- plot2 = False # plot additional results
- if plot2:
- ax = plt.subplots(2, 4, figsize=(10, 6), tight_layout=True)[1].ravel()
-
- fig2, ax2 = plt.subplots(1, 1, figsize=(8, 4), tight_layout=True)
- # for f in [save_dir / f'study_coco_{x}.txt' for x in ['yolov5n6', 'yolov5s6', 'yolov5m6', 'yolov5l6', 'yolov5x6']]:
- for f in sorted(save_dir.glob('study*.txt')):
- y = np.loadtxt(f, dtype=np.float32, usecols=[0, 1, 2, 3, 7, 8, 9], ndmin=2).T
- x = np.arange(y.shape[1]) if x is None else np.array(x)
- if plot2:
- s = ['P', 'R', 'mAP@.5', 'mAP@.5:.95', 't_preprocess (ms/img)', 't_inference (ms/img)', 't_NMS (ms/img)']
- for i in range(7):
- ax[i].plot(x, y[i], '.-', linewidth=2, markersize=8)
- ax[i].set_title(s[i])
-
- j = y[3].argmax() + 1
- ax2.plot(y[5, 1:j],
- y[3, 1:j] * 1E2,
- '.-',
- linewidth=2,
- markersize=8,
- label=f.stem.replace('study_coco_', '').replace('yolo', 'YOLO'))
-
- ax2.plot(1E3 / np.array([209, 140, 97, 58, 35, 18]), [34.6, 40.5, 43.0, 47.5, 49.7, 51.5],
- 'k.-',
- linewidth=2,
- markersize=8,
- alpha=.25,
- label='EfficientDet')
-
- ax2.grid(alpha=0.2)
- ax2.set_yticks(np.arange(20, 60, 5))
- ax2.set_xlim(0, 57)
- ax2.set_ylim(25, 55)
- ax2.set_xlabel('GPU Speed (ms/img)')
- ax2.set_ylabel('COCO AP val')
- ax2.legend(loc='lower right')
- f = save_dir / 'study.png'
- print(f'Saving {f}...')
- plt.savefig(f, dpi=300)
-
-
-@try_except # known issue https://github.com/ultralytics/yolov5/issues/5395
-@Timeout(30) # known issue https://github.com/ultralytics/yolov5/issues/5611
-def plot_labels(labels, names=(), save_dir=Path('')):
- # plot dataset labels
- LOGGER.info(f"Plotting labels to {save_dir / 'labels.jpg'}... ")
- c, b = labels[:, 0], labels[:, 1:].transpose() # classes, boxes
- nc = int(c.max() + 1) # number of classes
- x = pd.DataFrame(b.transpose(), columns=['x', 'y', 'width', 'height'])
-
- # seaborn correlogram
- sn.pairplot(x, corner=True, diag_kind='auto', kind='hist', diag_kws=dict(bins=50), plot_kws=dict(pmax=0.9))
- plt.savefig(save_dir / 'labels_correlogram.jpg', dpi=200)
- plt.close()
-
- # matplotlib labels
- matplotlib.use('svg') # faster
- ax = plt.subplots(2, 2, figsize=(8, 8), tight_layout=True)[1].ravel()
- y = ax[0].hist(c, bins=np.linspace(0, nc, nc + 1) - 0.5, rwidth=0.8)
- try: # color histogram bars by class
- [y[2].patches[i].set_color([x / 255 for x in colors(i)]) for i in range(nc)] # known issue #3195
- except Exception:
- pass
- ax[0].set_ylabel('instances')
- if 0 < len(names) < 30:
- ax[0].set_xticks(range(len(names)))
- ax[0].set_xticklabels(names, rotation=90, fontsize=10)
- else:
- ax[0].set_xlabel('classes')
- sn.histplot(x, x='x', y='y', ax=ax[2], bins=50, pmax=0.9)
- sn.histplot(x, x='width', y='height', ax=ax[3], bins=50, pmax=0.9)
-
- # rectangles
- labels[:, 1:3] = 0.5 # center
- labels[:, 1:] = xywh2xyxy(labels[:, 1:]) * 2000
- img = Image.fromarray(np.ones((2000, 2000, 3), dtype=np.uint8) * 255)
- for cls, *box in labels[:1000]:
- ImageDraw.Draw(img).rectangle(box, width=1, outline=colors(cls)) # plot
- ax[1].imshow(img)
- ax[1].axis('off')
-
- for a in [0, 1, 2, 3]:
- for s in ['top', 'right', 'left', 'bottom']:
- ax[a].spines[s].set_visible(False)
-
- plt.savefig(save_dir / 'labels.jpg', dpi=200)
- matplotlib.use('Agg')
- plt.close()
-
-
-def plot_evolve(evolve_csv='path/to/evolve.csv'): # from utils.plots import *; plot_evolve()
- # Plot evolve.csv hyp evolution results
- evolve_csv = Path(evolve_csv)
- data = pd.read_csv(evolve_csv)
- keys = [x.strip() for x in data.columns]
- x = data.values
- f = fitness(x)
- j = np.argmax(f) # max fitness index
- plt.figure(figsize=(10, 12), tight_layout=True)
- matplotlib.rc('font', **{'size': 8})
- print(f'Best results from row {j} of {evolve_csv}:')
- for i, k in enumerate(keys[7:]):
- v = x[:, 7 + i]
- mu = v[j] # best single result
- plt.subplot(6, 5, i + 1)
- plt.scatter(v, f, c=hist2d(v, f, 20), cmap='viridis', alpha=.8, edgecolors='none')
- plt.plot(mu, f.max(), 'k+', markersize=15)
- plt.title(f'{k} = {mu:.3g}', fontdict={'size': 9}) # limit to 40 characters
- if i % 5 != 0:
- plt.yticks([])
- print(f'{k:>15}: {mu:.3g}')
- f = evolve_csv.with_suffix('.png') # filename
- plt.savefig(f, dpi=200)
- plt.close()
- print(f'Saved {f}')
-
-
-def plot_results(file='path/to/results.csv', dir=''):
- # Plot training results.csv. Usage: from utils.plots import *; plot_results('path/to/results.csv')
- save_dir = Path(file).parent if file else Path(dir)
- fig, ax = plt.subplots(2, 5, figsize=(12, 6), tight_layout=True)
- ax = ax.ravel()
- files = list(save_dir.glob('results*.csv'))
- assert len(files), f'No results.csv files found in {save_dir.resolve()}, nothing to plot.'
- for f in files:
- try:
- data = pd.read_csv(f)
- s = [x.strip() for x in data.columns]
- x = data.values[:, 0]
- for i, j in enumerate([1, 2, 3, 4, 5, 8, 9, 10, 6, 7]):
- y = data.values[:, j].astype('float')
- # y[y == 0] = np.nan # don't show zero values
- ax[i].plot(x, y, marker='.', label=f.stem, linewidth=2, markersize=8)
- ax[i].set_title(s[j], fontsize=12)
- # if j in [8, 9, 10]: # share train and val loss y axes
- # ax[i].get_shared_y_axes().join(ax[i], ax[i - 5])
- except Exception as e:
- LOGGER.info(f'Warning: Plotting error for {f}: {e}')
- ax[1].legend()
- fig.savefig(save_dir / 'results.png', dpi=200)
- plt.close()
-
-
-def profile_idetection(start=0, stop=0, labels=(), save_dir=''):
- # Plot iDetection '*.txt' per-image logs. from utils.plots import *; profile_idetection()
- ax = plt.subplots(2, 4, figsize=(12, 6), tight_layout=True)[1].ravel()
- s = ['Images', 'Free Storage (GB)', 'RAM Usage (GB)', 'Battery', 'dt_raw (ms)', 'dt_smooth (ms)', 'real-world FPS']
- files = list(Path(save_dir).glob('frames*.txt'))
- for fi, f in enumerate(files):
- try:
- results = np.loadtxt(f, ndmin=2).T[:, 90:-30] # clip first and last rows
- n = results.shape[1] # number of rows
- x = np.arange(start, min(stop, n) if stop else n)
- results = results[:, x]
- t = (results[0] - results[0].min()) # set t0=0s
- results[0] = x
- for i, a in enumerate(ax):
- if i < len(results):
- label = labels[fi] if len(labels) else f.stem.replace('frames_', '')
- a.plot(t, results[i], marker='.', label=label, linewidth=1, markersize=5)
- a.set_title(s[i])
- a.set_xlabel('time (s)')
- # if fi == len(files) - 1:
- # a.set_ylim(bottom=0)
- for side in ['top', 'right']:
- a.spines[side].set_visible(False)
- else:
- a.remove()
- except Exception as e:
- print(f'Warning: Plotting error for {f}; {e}')
- ax[1].legend()
- plt.savefig(Path(save_dir) / 'idetection_profile.png', dpi=200)
-
-
-def save_one_box(xyxy, im, file=Path('im.jpg'), gain=1.02, pad=10, square=False, BGR=False, save=True):
- # Save image crop as {file} with crop size multiple {gain} and {pad} pixels. Save and/or return crop
- xyxy = torch.tensor(xyxy).view(-1, 4)
- b = xyxy2xywh(xyxy) # boxes
- if square:
- b[:, 2:] = b[:, 2:].max(1)[0].unsqueeze(1) # attempt rectangle to square
- b[:, 2:] = b[:, 2:] * gain + pad # box wh * gain + pad
- xyxy = xywh2xyxy(b).long()
- clip_coords(xyxy, im.shape)
- crop = im[int(xyxy[0, 1]):int(xyxy[0, 3]), int(xyxy[0, 0]):int(xyxy[0, 2]), ::(1 if BGR else -1)]
- if save:
- file.parent.mkdir(parents=True, exist_ok=True) # make directory
- f = str(increment_path(file).with_suffix('.jpg'))
- # cv2.imwrite(f, crop) # https://github.com/ultralytics/yolov5/issues/7007 chroma subsampling issue
- Image.fromarray(cv2.cvtColor(crop, cv2.COLOR_BGR2RGB)).save(f, quality=95, subsampling=0)
- return crop
diff --git a/spaces/caffeinum/VToonify/vtoonify/model/raft/train.py b/spaces/caffeinum/VToonify/vtoonify/model/raft/train.py
deleted file mode 100644
index 307573097f13ee30c67bbe11658f457fdf1ead3c..0000000000000000000000000000000000000000
--- a/spaces/caffeinum/VToonify/vtoonify/model/raft/train.py
+++ /dev/null
@@ -1,247 +0,0 @@
-from __future__ import print_function, division
-import sys
-sys.path.append('core')
-
-import argparse
-import os
-import cv2
-import time
-import numpy as np
-import matplotlib.pyplot as plt
-
-import torch
-import torch.nn as nn
-import torch.optim as optim
-import torch.nn.functional as F
-
-from torch.utils.data import DataLoader
-from raft import RAFT
-import evaluate
-import datasets
-
-from torch.utils.tensorboard import SummaryWriter
-
-try:
- from torch.cuda.amp import GradScaler
-except:
- # dummy GradScaler for PyTorch < 1.6
- class GradScaler:
- def __init__(self):
- pass
- def scale(self, loss):
- return loss
- def unscale_(self, optimizer):
- pass
- def step(self, optimizer):
- optimizer.step()
- def update(self):
- pass
-
-
-# exclude extremly large displacements
-MAX_FLOW = 400
-SUM_FREQ = 100
-VAL_FREQ = 5000
-
-
-def sequence_loss(flow_preds, flow_gt, valid, gamma=0.8, max_flow=MAX_FLOW):
- """ Loss function defined over sequence of flow predictions """
-
- n_predictions = len(flow_preds)
- flow_loss = 0.0
-
- # exlude invalid pixels and extremely large diplacements
- mag = torch.sum(flow_gt**2, dim=1).sqrt()
- valid = (valid >= 0.5) & (mag < max_flow)
-
- for i in range(n_predictions):
- i_weight = gamma**(n_predictions - i - 1)
- i_loss = (flow_preds[i] - flow_gt).abs()
- flow_loss += i_weight * (valid[:, None] * i_loss).mean()
-
- epe = torch.sum((flow_preds[-1] - flow_gt)**2, dim=1).sqrt()
- epe = epe.view(-1)[valid.view(-1)]
-
- metrics = {
- 'epe': epe.mean().item(),
- '1px': (epe < 1).float().mean().item(),
- '3px': (epe < 3).float().mean().item(),
- '5px': (epe < 5).float().mean().item(),
- }
-
- return flow_loss, metrics
-
-
-def count_parameters(model):
- return sum(p.numel() for p in model.parameters() if p.requires_grad)
-
-
-def fetch_optimizer(args, model):
- """ Create the optimizer and learning rate scheduler """
- optimizer = optim.AdamW(model.parameters(), lr=args.lr, weight_decay=args.wdecay, eps=args.epsilon)
-
- scheduler = optim.lr_scheduler.OneCycleLR(optimizer, args.lr, args.num_steps+100,
- pct_start=0.05, cycle_momentum=False, anneal_strategy='linear')
-
- return optimizer, scheduler
-
-
-class Logger:
- def __init__(self, model, scheduler):
- self.model = model
- self.scheduler = scheduler
- self.total_steps = 0
- self.running_loss = {}
- self.writer = None
-
- def _print_training_status(self):
- metrics_data = [self.running_loss[k]/SUM_FREQ for k in sorted(self.running_loss.keys())]
- training_str = "[{:6d}, {:10.7f}] ".format(self.total_steps+1, self.scheduler.get_last_lr()[0])
- metrics_str = ("{:10.4f}, "*len(metrics_data)).format(*metrics_data)
-
- # print the training status
- print(training_str + metrics_str)
-
- if self.writer is None:
- self.writer = SummaryWriter()
-
- for k in self.running_loss:
- self.writer.add_scalar(k, self.running_loss[k]/SUM_FREQ, self.total_steps)
- self.running_loss[k] = 0.0
-
- def push(self, metrics):
- self.total_steps += 1
-
- for key in metrics:
- if key not in self.running_loss:
- self.running_loss[key] = 0.0
-
- self.running_loss[key] += metrics[key]
-
- if self.total_steps % SUM_FREQ == SUM_FREQ-1:
- self._print_training_status()
- self.running_loss = {}
-
- def write_dict(self, results):
- if self.writer is None:
- self.writer = SummaryWriter()
-
- for key in results:
- self.writer.add_scalar(key, results[key], self.total_steps)
-
- def close(self):
- self.writer.close()
-
-
-def train(args):
-
- model = nn.DataParallel(RAFT(args), device_ids=args.gpus)
- print("Parameter Count: %d" % count_parameters(model))
-
- if args.restore_ckpt is not None:
- model.load_state_dict(torch.load(args.restore_ckpt), strict=False)
-
- model.cuda()
- model.train()
-
- if args.stage != 'chairs':
- model.module.freeze_bn()
-
- train_loader = datasets.fetch_dataloader(args)
- optimizer, scheduler = fetch_optimizer(args, model)
-
- total_steps = 0
- scaler = GradScaler(enabled=args.mixed_precision)
- logger = Logger(model, scheduler)
-
- VAL_FREQ = 5000
- add_noise = True
-
- should_keep_training = True
- while should_keep_training:
-
- for i_batch, data_blob in enumerate(train_loader):
- optimizer.zero_grad()
- image1, image2, flow, valid = [x.cuda() for x in data_blob]
-
- if args.add_noise:
- stdv = np.random.uniform(0.0, 5.0)
- image1 = (image1 + stdv * torch.randn(*image1.shape).cuda()).clamp(0.0, 255.0)
- image2 = (image2 + stdv * torch.randn(*image2.shape).cuda()).clamp(0.0, 255.0)
-
- flow_predictions = model(image1, image2, iters=args.iters)
-
- loss, metrics = sequence_loss(flow_predictions, flow, valid, args.gamma)
- scaler.scale(loss).backward()
- scaler.unscale_(optimizer)
- torch.nn.utils.clip_grad_norm_(model.parameters(), args.clip)
-
- scaler.step(optimizer)
- scheduler.step()
- scaler.update()
-
- logger.push(metrics)
-
- if total_steps % VAL_FREQ == VAL_FREQ - 1:
- PATH = 'checkpoints/%d_%s.pth' % (total_steps+1, args.name)
- torch.save(model.state_dict(), PATH)
-
- results = {}
- for val_dataset in args.validation:
- if val_dataset == 'chairs':
- results.update(evaluate.validate_chairs(model.module))
- elif val_dataset == 'sintel':
- results.update(evaluate.validate_sintel(model.module))
- elif val_dataset == 'kitti':
- results.update(evaluate.validate_kitti(model.module))
-
- logger.write_dict(results)
-
- model.train()
- if args.stage != 'chairs':
- model.module.freeze_bn()
-
- total_steps += 1
-
- if total_steps > args.num_steps:
- should_keep_training = False
- break
-
- logger.close()
- PATH = 'checkpoints/%s.pth' % args.name
- torch.save(model.state_dict(), PATH)
-
- return PATH
-
-
-if __name__ == '__main__':
- parser = argparse.ArgumentParser()
- parser.add_argument('--name', default='raft', help="name your experiment")
- parser.add_argument('--stage', help="determines which dataset to use for training")
- parser.add_argument('--restore_ckpt', help="restore checkpoint")
- parser.add_argument('--small', action='store_true', help='use small model')
- parser.add_argument('--validation', type=str, nargs='+')
-
- parser.add_argument('--lr', type=float, default=0.00002)
- parser.add_argument('--num_steps', type=int, default=100000)
- parser.add_argument('--batch_size', type=int, default=6)
- parser.add_argument('--image_size', type=int, nargs='+', default=[384, 512])
- parser.add_argument('--gpus', type=int, nargs='+', default=[0,1])
- parser.add_argument('--mixed_precision', action='store_true', help='use mixed precision')
-
- parser.add_argument('--iters', type=int, default=12)
- parser.add_argument('--wdecay', type=float, default=.00005)
- parser.add_argument('--epsilon', type=float, default=1e-8)
- parser.add_argument('--clip', type=float, default=1.0)
- parser.add_argument('--dropout', type=float, default=0.0)
- parser.add_argument('--gamma', type=float, default=0.8, help='exponential weighting')
- parser.add_argument('--add_noise', action='store_true')
- args = parser.parse_args()
-
- torch.manual_seed(1234)
- np.random.seed(1234)
-
- if not os.path.isdir('checkpoints'):
- os.mkdir('checkpoints')
-
- train(args)
\ No newline at end of file
diff --git a/spaces/carlosalonso/Detection-video/carpeta_deteccion/detectron2/layers/aspp.py b/spaces/carlosalonso/Detection-video/carpeta_deteccion/detectron2/layers/aspp.py
deleted file mode 100644
index 14861aa9ede4fea6a69a49f189bcab997b558148..0000000000000000000000000000000000000000
--- a/spaces/carlosalonso/Detection-video/carpeta_deteccion/detectron2/layers/aspp.py
+++ /dev/null
@@ -1,144 +0,0 @@
-# Copyright (c) Facebook, Inc. and its affiliates.
-
-from copy import deepcopy
-import fvcore.nn.weight_init as weight_init
-import torch
-from torch import nn
-from torch.nn import functional as F
-
-from .batch_norm import get_norm
-from .blocks import DepthwiseSeparableConv2d
-from .wrappers import Conv2d
-
-
-class ASPP(nn.Module):
- """
- Atrous Spatial Pyramid Pooling (ASPP).
- """
-
- def __init__(
- self,
- in_channels,
- out_channels,
- dilations,
- *,
- norm,
- activation,
- pool_kernel_size=None,
- dropout: float = 0.0,
- use_depthwise_separable_conv=False,
- ):
- """
- Args:
- in_channels (int): number of input channels for ASPP.
- out_channels (int): number of output channels.
- dilations (list): a list of 3 dilations in ASPP.
- norm (str or callable): normalization for all conv layers.
- See :func:`layers.get_norm` for supported format. norm is
- applied to all conv layers except the conv following
- global average pooling.
- activation (callable): activation function.
- pool_kernel_size (tuple, list): the average pooling size (kh, kw)
- for image pooling layer in ASPP. If set to None, it always
- performs global average pooling. If not None, it must be
- divisible by the shape of inputs in forward(). It is recommended
- to use a fixed input feature size in training, and set this
- option to match this size, so that it performs global average
- pooling in training, and the size of the pooling window stays
- consistent in inference.
- dropout (float): apply dropout on the output of ASPP. It is used in
- the official DeepLab implementation with a rate of 0.1:
- https://github.com/tensorflow/models/blob/21b73d22f3ed05b650e85ac50849408dd36de32e/research/deeplab/model.py#L532 # noqa
- use_depthwise_separable_conv (bool): use DepthwiseSeparableConv2d
- for 3x3 convs in ASPP, proposed in :paper:`DeepLabV3+`.
- """
- super(ASPP, self).__init__()
- assert len(dilations) == 3, "ASPP expects 3 dilations, got {}".format(len(dilations))
- self.pool_kernel_size = pool_kernel_size
- self.dropout = dropout
- use_bias = norm == ""
- self.convs = nn.ModuleList()
- # conv 1x1
- self.convs.append(
- Conv2d(
- in_channels,
- out_channels,
- kernel_size=1,
- bias=use_bias,
- norm=get_norm(norm, out_channels),
- activation=deepcopy(activation),
- )
- )
- weight_init.c2_xavier_fill(self.convs[-1])
- # atrous convs
- for dilation in dilations:
- if use_depthwise_separable_conv:
- self.convs.append(
- DepthwiseSeparableConv2d(
- in_channels,
- out_channels,
- kernel_size=3,
- padding=dilation,
- dilation=dilation,
- norm1=norm,
- activation1=deepcopy(activation),
- norm2=norm,
- activation2=deepcopy(activation),
- )
- )
- else:
- self.convs.append(
- Conv2d(
- in_channels,
- out_channels,
- kernel_size=3,
- padding=dilation,
- dilation=dilation,
- bias=use_bias,
- norm=get_norm(norm, out_channels),
- activation=deepcopy(activation),
- )
- )
- weight_init.c2_xavier_fill(self.convs[-1])
- # image pooling
- # We do not add BatchNorm because the spatial resolution is 1x1,
- # the original TF implementation has BatchNorm.
- if pool_kernel_size is None:
- image_pooling = nn.Sequential(
- nn.AdaptiveAvgPool2d(1),
- Conv2d(in_channels, out_channels, 1, bias=True, activation=deepcopy(activation)),
- )
- else:
- image_pooling = nn.Sequential(
- nn.AvgPool2d(kernel_size=pool_kernel_size, stride=1),
- Conv2d(in_channels, out_channels, 1, bias=True, activation=deepcopy(activation)),
- )
- weight_init.c2_xavier_fill(image_pooling[1])
- self.convs.append(image_pooling)
-
- self.project = Conv2d(
- 5 * out_channels,
- out_channels,
- kernel_size=1,
- bias=use_bias,
- norm=get_norm(norm, out_channels),
- activation=deepcopy(activation),
- )
- weight_init.c2_xavier_fill(self.project)
-
- def forward(self, x):
- size = x.shape[-2:]
- if self.pool_kernel_size is not None:
- if size[0] % self.pool_kernel_size[0] or size[1] % self.pool_kernel_size[1]:
- raise ValueError(
- "`pool_kernel_size` must be divisible by the shape of inputs. "
- "Input size: {} `pool_kernel_size`: {}".format(size, self.pool_kernel_size)
- )
- res = []
- for conv in self.convs:
- res.append(conv(x))
- res[-1] = F.interpolate(res[-1], size=size, mode="bilinear", align_corners=False)
- res = torch.cat(res, dim=1)
- res = self.project(res)
- res = F.dropout(res, self.dropout, training=self.training) if self.dropout > 0 else res
- return res
diff --git a/spaces/carlosalonso/Detection-video/carpeta_deteccion/tools/lazyconfig_train_net.py b/spaces/carlosalonso/Detection-video/carpeta_deteccion/tools/lazyconfig_train_net.py
deleted file mode 100644
index bb62d36c0c171b0391453afafc2828ebab1b0da1..0000000000000000000000000000000000000000
--- a/spaces/carlosalonso/Detection-video/carpeta_deteccion/tools/lazyconfig_train_net.py
+++ /dev/null
@@ -1,131 +0,0 @@
-#!/usr/bin/env python
-# Copyright (c) Facebook, Inc. and its affiliates.
-"""
-Training script using the new "LazyConfig" python config files.
-
-This scripts reads a given python config file and runs the training or evaluation.
-It can be used to train any models or dataset as long as they can be
-instantiated by the recursive construction defined in the given config file.
-
-Besides lazy construction of models, dataloader, etc., this scripts expects a
-few common configuration parameters currently defined in "configs/common/train.py".
-To add more complicated training logic, you can easily add other configs
-in the config file and implement a new train_net.py to handle them.
-"""
-import logging
-
-from detectron2.checkpoint import DetectionCheckpointer
-from detectron2.config import LazyConfig, instantiate
-from detectron2.engine import (
- AMPTrainer,
- SimpleTrainer,
- default_argument_parser,
- default_setup,
- default_writers,
- hooks,
- launch,
-)
-from detectron2.engine.defaults import create_ddp_model
-from detectron2.evaluation import inference_on_dataset, print_csv_format
-from detectron2.utils import comm
-
-logger = logging.getLogger("detectron2")
-
-
-def do_test(cfg, model):
- if "evaluator" in cfg.dataloader:
- ret = inference_on_dataset(
- model, instantiate(cfg.dataloader.test), instantiate(cfg.dataloader.evaluator)
- )
- print_csv_format(ret)
- return ret
-
-
-def do_train(args, cfg):
- """
- Args:
- cfg: an object with the following attributes:
- model: instantiate to a module
- dataloader.{train,test}: instantiate to dataloaders
- dataloader.evaluator: instantiate to evaluator for test set
- optimizer: instantaite to an optimizer
- lr_multiplier: instantiate to a fvcore scheduler
- train: other misc config defined in `configs/common/train.py`, including:
- output_dir (str)
- init_checkpoint (str)
- amp.enabled (bool)
- max_iter (int)
- eval_period, log_period (int)
- device (str)
- checkpointer (dict)
- ddp (dict)
- """
- model = instantiate(cfg.model)
- logger = logging.getLogger("detectron2")
- logger.info("Model:\n{}".format(model))
- model.to(cfg.train.device)
-
- cfg.optimizer.params.model = model
- optim = instantiate(cfg.optimizer)
-
- train_loader = instantiate(cfg.dataloader.train)
-
- model = create_ddp_model(model, **cfg.train.ddp)
- trainer = (AMPTrainer if cfg.train.amp.enabled else SimpleTrainer)(model, train_loader, optim)
- checkpointer = DetectionCheckpointer(
- model,
- cfg.train.output_dir,
- trainer=trainer,
- )
- trainer.register_hooks(
- [
- hooks.IterationTimer(),
- hooks.LRScheduler(scheduler=instantiate(cfg.lr_multiplier)),
- hooks.PeriodicCheckpointer(checkpointer, **cfg.train.checkpointer)
- if comm.is_main_process()
- else None,
- hooks.EvalHook(cfg.train.eval_period, lambda: do_test(cfg, model)),
- hooks.PeriodicWriter(
- default_writers(cfg.train.output_dir, cfg.train.max_iter),
- period=cfg.train.log_period,
- )
- if comm.is_main_process()
- else None,
- ]
- )
-
- checkpointer.resume_or_load(cfg.train.init_checkpoint, resume=args.resume)
- if args.resume and checkpointer.has_checkpoint():
- # The checkpoint stores the training iteration that just finished, thus we start
- # at the next iteration
- start_iter = trainer.iter + 1
- else:
- start_iter = 0
- trainer.train(start_iter, cfg.train.max_iter)
-
-
-def main(args):
- cfg = LazyConfig.load(args.config_file)
- cfg = LazyConfig.apply_overrides(cfg, args.opts)
- default_setup(cfg, args)
-
- if args.eval_only:
- model = instantiate(cfg.model)
- model.to(cfg.train.device)
- model = create_ddp_model(model)
- DetectionCheckpointer(model).load(cfg.train.init_checkpoint)
- print(do_test(cfg, model))
- else:
- do_train(args, cfg)
-
-
-if __name__ == "__main__":
- args = default_argument_parser().parse_args()
- launch(
- main,
- args.num_gpus,
- num_machines=args.num_machines,
- machine_rank=args.machine_rank,
- dist_url=args.dist_url,
- args=(args,),
- )
diff --git a/spaces/chansung/palm-with-gradio-chat/js.py b/spaces/chansung/palm-with-gradio-chat/js.py
deleted file mode 100644
index 781e4c35f98903536b1fcdb075a331988698eeb9..0000000000000000000000000000000000000000
--- a/spaces/chansung/palm-with-gradio-chat/js.py
+++ /dev/null
@@ -1,81 +0,0 @@
-GET_LOCAL_STORAGE = """
-function() {
- globalThis.setStorage = (key, value)=>{
- localStorage.setItem(key, JSON.stringify(value));
- }
- globalThis.getStorage = (key, value)=>{
- return JSON.parse(localStorage.getItem(key));
- }
-
- var local_data = getStorage('local_data');
- var history = [];
-
- if(local_data) {
- local_data[0].pingpongs.forEach(element =>{
- history.push([element.ping, element.pong]);
- });
- }
- else {
- local_data = [];
- for (let step = 0; step < 10; step++) {
- local_data.push({'ctx': '', 'pingpongs':[]});
- }
- setStorage('local_data', local_data);
- }
-
- if(history.length == 0) {
- document.querySelector("#initial-popup").classList.remove('hide');
- }
-
- return [history, local_data];
-}
-"""
-
-UPDATE_LEFT_BTNS_STATE = """
-(v)=>{
- document.querySelector('.custom-btn-highlight').classList.add('custom-btn');
- document.querySelector('.custom-btn-highlight').classList.remove('custom-btn-highlight');
-
- const elements = document.querySelectorAll(".custom-btn");
-
- for(var i=0; i < elements.length; i++) {
- const element = elements[i];
- if(element.textContent == v) {
- console.log(v);
- element.classList.add('custom-btn-highlight');
- element.classList.remove('custom-btn');
- break;
- }
- }
-}"""
-
-UPDATE_PLACEHOLDERS = """
-function update_placeholders(txt, placeholder_txt1, placeholder_txt2, placeholder_txt3) {
- let example_prompt = txt;
-
- const regex = /\[([^\]]*)\]/g;
- const matches = txt.match(regex);
-
- if (matches != null) {
- if (matches.length >= 1) {
- if (placeholder_txt1 !== "") {
- example_prompt = example_prompt.replace(matches[0], placeholder_txt1);
- }
- }
-
- if (matches.length >= 2) {
- if (placeholder_txt2 !== "") {
- example_prompt = example_prompt.replace(matches[1], placeholder_txt2);
- }
- }
-
- if (matches.length >= 3) {
- if (placeholder_txt1 !== "") {
- example_prompt = example_prompt.replace(matches[2], placeholder_txt3);
- }
- }
- }
-
- return example_prompt
-}
-"""
\ No newline at end of file
diff --git a/spaces/chats-bug/ai-image-captioning/app.py b/spaces/chats-bug/ai-image-captioning/app.py
deleted file mode 100644
index 180841c336970ac66d8ad292095a6b5887498a42..0000000000000000000000000000000000000000
--- a/spaces/chats-bug/ai-image-captioning/app.py
+++ /dev/null
@@ -1,102 +0,0 @@
-import gradio as gr
-import torch
-from PIL import Image
-
-from model import BlipBaseModel, GitBaseCocoModel
-
-MODELS = {
- "Git-Base-COCO": GitBaseCocoModel,
- "Blip Base": BlipBaseModel,
-}
-
-# examples = [["Image1.png"], ["Image2.png"], ["Image3.png"]]
-
-def generate_captions(
- image,
- num_captions,
- model_name,
- max_length,
- temperature,
- top_k,
- top_p,
- repetition_penalty,
- diversity_penalty,
- ):
- """
- Generates captions for the given image.
-
- -----
- Parameters:
- image: PIL.Image
- The image to generate captions for.
- num_captions: int
- The number of captions to generate.
- ** Rest of the parameters are the same as in the model.generate method. **
- -----
- Returns:
- list[str]
- """
- # Convert the numerical values to their corresponding types.
- # Gradio Slider returns values as floats: except when the value is a whole number, in which case it returns an int.
- # Only float values suffer from this issue.
- temperature = float(temperature)
- top_p = float(top_p)
- repetition_penalty = float(repetition_penalty)
- diversity_penalty = float(diversity_penalty)
-
- device = "cuda" if torch.cuda.is_available() else "cpu"
-
- model = MODELS[model_name](device)
-
- captions = model.generate(
- image=image,
- max_length=max_length,
- num_captions=num_captions,
- temperature=temperature,
- top_k=top_k,
- top_p=top_p,
- repetition_penalty=repetition_penalty,
- diversity_penalty=diversity_penalty,
- )
-
- # Convert list to a single string separated by newlines.
- captions = "\n".join(captions)
- return captions
-
-title = "AI tool for generating captions for images"
-description = "This tool uses pretrained models to generate captions for images."
-
-interface = gr.Interface(
- fn=generate_captions,
- inputs=[
- gr.components.Image(type="pil", label="Image"),
- gr.components.Slider(minimum=1, maximum=10, step=1, value=1, label="Number of Captions to Generate"),
- gr.components.Dropdown(MODELS.keys(), label="Model", value=list(MODELS.keys())[1]), # Default to Blip Base
- gr.components.Slider(minimum=20, maximum=100, step=5, value=50, label="Maximum Caption Length"),
- gr.components.Slider(minimum=0.1, maximum=10.0, step=0.1, value=1.0, label="Temperature"),
- gr.components.Slider(minimum=1, maximum=100, step=1, value=50, label="Top K"),
- gr.components.Slider(minimum=0.1, maximum=5.0, step=0.1, value=1.0, label="Top P"),
- gr.components.Slider(minimum=1.0, maximum=10.0, step=0.1, value=2.0, label="Repetition Penalty"),
- gr.components.Slider(minimum=0.0, maximum=10.0, step=0.1, value=2.0, label="Diversity Penalty"),
- ],
- outputs=[
- gr.components.Textbox(label="Caption"),
- ],
- # Set image examples to be displayed in the interface.
- examples = [
- ["Image1.png", 1, list(MODELS.keys())[1], 50, 1.0, 50, 1.0, 2.0, 2.0],
- ["Image2.png", 1, list(MODELS.keys())[1], 50, 1.0, 50, 1.0, 2.0, 2.0],
- ["Image3.png", 1, list(MODELS.keys())[1], 50, 1.0, 50, 1.0, 2.0, 2.0],
- ],
- title=title,
- description=description,
- allow_flagging="never",
-)
-
-
-if __name__ == "__main__":
- # Launch the interface.
- interface.launch(
- enable_queue=True,
- debug=True,
- )
\ No newline at end of file
diff --git a/spaces/christhegamechanger/background_swapping/setup.sh b/spaces/christhegamechanger/background_swapping/setup.sh
deleted file mode 100644
index c8650a8b74a58d9a5f53b185fd711c5668e1cd52..0000000000000000000000000000000000000000
--- a/spaces/christhegamechanger/background_swapping/setup.sh
+++ /dev/null
@@ -1,13 +0,0 @@
-mkdir -p ~/.streamlit/
-
-echo "\
-[general]\n\
-email = \"your-email@domain.com\"\n\
-" > ~/.streamlit/credentials.toml
-
-echo "\
-[server]\n\
-headless = true\n\
-enableCORS=false\n\
-port = $PORT\n\
-" > ~/.streamlit/config.toml
\ No newline at end of file
diff --git a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/fontTools/ttLib/ttGlyphSet.py b/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/fontTools/ttLib/ttGlyphSet.py
deleted file mode 100644
index fa7fbd4f23558f6705ee3e819ded518bb7549e36..0000000000000000000000000000000000000000
--- a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/fontTools/ttLib/ttGlyphSet.py
+++ /dev/null
@@ -1,322 +0,0 @@
-"""GlyphSets returned by a TTFont."""
-
-from abc import ABC, abstractmethod
-from collections.abc import Mapping
-from contextlib import contextmanager
-from copy import copy
-from types import SimpleNamespace
-from fontTools.misc.fixedTools import otRound
-from fontTools.misc.loggingTools import deprecateFunction
-from fontTools.misc.transform import Transform
-from fontTools.pens.transformPen import TransformPen, TransformPointPen
-
-
-class _TTGlyphSet(Mapping):
-
- """Generic dict-like GlyphSet class that pulls metrics from hmtx and
- glyph shape from TrueType or CFF.
- """
-
- def __init__(self, font, location, glyphsMapping):
- self.font = font
- self.defaultLocationNormalized = (
- {axis.axisTag: 0 for axis in self.font["fvar"].axes}
- if "fvar" in self.font
- else {}
- )
- self.location = location if location is not None else {}
- self.rawLocation = {} # VarComponent-only location
- self.originalLocation = location if location is not None else {}
- self.depth = 0
- self.locationStack = []
- self.rawLocationStack = []
- self.glyphsMapping = glyphsMapping
- self.hMetrics = font["hmtx"].metrics
- self.vMetrics = getattr(font.get("vmtx"), "metrics", None)
- self.hvarTable = None
- if location:
- from fontTools.varLib.varStore import VarStoreInstancer
-
- self.hvarTable = getattr(font.get("HVAR"), "table", None)
- if self.hvarTable is not None:
- self.hvarInstancer = VarStoreInstancer(
- self.hvarTable.VarStore, font["fvar"].axes, location
- )
- # TODO VVAR, VORG
-
- @contextmanager
- def pushLocation(self, location, reset: bool):
- self.locationStack.append(self.location)
- self.rawLocationStack.append(self.rawLocation)
- if reset:
- self.location = self.originalLocation.copy()
- self.rawLocation = self.defaultLocationNormalized.copy()
- else:
- self.location = self.location.copy()
- self.rawLocation = {}
- self.location.update(location)
- self.rawLocation.update(location)
-
- try:
- yield None
- finally:
- self.location = self.locationStack.pop()
- self.rawLocation = self.rawLocationStack.pop()
-
- @contextmanager
- def pushDepth(self):
- try:
- depth = self.depth
- self.depth += 1
- yield depth
- finally:
- self.depth -= 1
-
- def __contains__(self, glyphName):
- return glyphName in self.glyphsMapping
-
- def __iter__(self):
- return iter(self.glyphsMapping.keys())
-
- def __len__(self):
- return len(self.glyphsMapping)
-
- @deprecateFunction(
- "use 'glyphName in glyphSet' instead", category=DeprecationWarning
- )
- def has_key(self, glyphName):
- return glyphName in self.glyphsMapping
-
-
-class _TTGlyphSetGlyf(_TTGlyphSet):
- def __init__(self, font, location):
- self.glyfTable = font["glyf"]
- super().__init__(font, location, self.glyfTable)
- self.gvarTable = font.get("gvar")
-
- def __getitem__(self, glyphName):
- return _TTGlyphGlyf(self, glyphName)
-
-
-class _TTGlyphSetCFF(_TTGlyphSet):
- def __init__(self, font, location):
- tableTag = "CFF2" if "CFF2" in font else "CFF "
- self.charStrings = list(font[tableTag].cff.values())[0].CharStrings
- super().__init__(font, location, self.charStrings)
- self.blender = None
- if location:
- from fontTools.varLib.varStore import VarStoreInstancer
-
- varStore = getattr(self.charStrings, "varStore", None)
- if varStore is not None:
- instancer = VarStoreInstancer(
- varStore.otVarStore, font["fvar"].axes, location
- )
- self.blender = instancer.interpolateFromDeltas
-
- def __getitem__(self, glyphName):
- return _TTGlyphCFF(self, glyphName)
-
-
-class _TTGlyph(ABC):
-
- """Glyph object that supports the Pen protocol, meaning that it has
- .draw() and .drawPoints() methods that take a pen object as their only
- argument. Additionally there are 'width' and 'lsb' attributes, read from
- the 'hmtx' table.
-
- If the font contains a 'vmtx' table, there will also be 'height' and 'tsb'
- attributes.
- """
-
- def __init__(self, glyphSet, glyphName):
- self.glyphSet = glyphSet
- self.name = glyphName
- self.width, self.lsb = glyphSet.hMetrics[glyphName]
- if glyphSet.vMetrics is not None:
- self.height, self.tsb = glyphSet.vMetrics[glyphName]
- else:
- self.height, self.tsb = None, None
- if glyphSet.location and glyphSet.hvarTable is not None:
- varidx = (
- glyphSet.font.getGlyphID(glyphName)
- if glyphSet.hvarTable.AdvWidthMap is None
- else glyphSet.hvarTable.AdvWidthMap.mapping[glyphName]
- )
- self.width += glyphSet.hvarInstancer[varidx]
- # TODO: VVAR/VORG
-
- @abstractmethod
- def draw(self, pen):
- """Draw the glyph onto ``pen``. See fontTools.pens.basePen for details
- how that works.
- """
- raise NotImplementedError
-
- def drawPoints(self, pen):
- """Draw the glyph onto ``pen``. See fontTools.pens.pointPen for details
- how that works.
- """
- from fontTools.pens.pointPen import SegmentToPointPen
-
- self.draw(SegmentToPointPen(pen))
-
-
-class _TTGlyphGlyf(_TTGlyph):
- def draw(self, pen):
- """Draw the glyph onto ``pen``. See fontTools.pens.basePen for details
- how that works.
- """
- glyph, offset = self._getGlyphAndOffset()
-
- with self.glyphSet.pushDepth() as depth:
-
- if depth:
- offset = 0 # Offset should only apply at top-level
-
- if glyph.isVarComposite():
- self._drawVarComposite(glyph, pen, False)
- return
-
- glyph.draw(pen, self.glyphSet.glyfTable, offset)
-
- def drawPoints(self, pen):
- """Draw the glyph onto ``pen``. See fontTools.pens.pointPen for details
- how that works.
- """
- glyph, offset = self._getGlyphAndOffset()
-
- with self.glyphSet.pushDepth() as depth:
-
- if depth:
- offset = 0 # Offset should only apply at top-level
-
- if glyph.isVarComposite():
- self._drawVarComposite(glyph, pen, True)
- return
-
- glyph.drawPoints(pen, self.glyphSet.glyfTable, offset)
-
- def _drawVarComposite(self, glyph, pen, isPointPen):
-
- from fontTools.ttLib.tables._g_l_y_f import (
- VarComponentFlags,
- VAR_COMPONENT_TRANSFORM_MAPPING,
- )
-
- for comp in glyph.components:
-
- with self.glyphSet.pushLocation(
- comp.location, comp.flags & VarComponentFlags.RESET_UNSPECIFIED_AXES
- ):
- try:
- pen.addVarComponent(
- comp.glyphName, comp.transform, self.glyphSet.rawLocation
- )
- except AttributeError:
- t = comp.transform.toTransform()
- if isPointPen:
- tPen = TransformPointPen(pen, t)
- self.glyphSet[comp.glyphName].drawPoints(tPen)
- else:
- tPen = TransformPen(pen, t)
- self.glyphSet[comp.glyphName].draw(tPen)
-
- def _getGlyphAndOffset(self):
- if self.glyphSet.location and self.glyphSet.gvarTable is not None:
- glyph = self._getGlyphInstance()
- else:
- glyph = self.glyphSet.glyfTable[self.name]
-
- offset = self.lsb - glyph.xMin if hasattr(glyph, "xMin") else 0
- return glyph, offset
-
- def _getGlyphInstance(self):
- from fontTools.varLib.iup import iup_delta
- from fontTools.ttLib.tables._g_l_y_f import GlyphCoordinates
- from fontTools.varLib.models import supportScalar
-
- glyphSet = self.glyphSet
- glyfTable = glyphSet.glyfTable
- variations = glyphSet.gvarTable.variations[self.name]
- hMetrics = glyphSet.hMetrics
- vMetrics = glyphSet.vMetrics
- coordinates, _ = glyfTable._getCoordinatesAndControls(
- self.name, hMetrics, vMetrics
- )
- origCoords, endPts = None, None
- for var in variations:
- scalar = supportScalar(glyphSet.location, var.axes)
- if not scalar:
- continue
- delta = var.coordinates
- if None in delta:
- if origCoords is None:
- origCoords, control = glyfTable._getCoordinatesAndControls(
- self.name, hMetrics, vMetrics
- )
- endPts = (
- control[1] if control[0] >= 1 else list(range(len(control[1])))
- )
- delta = iup_delta(delta, origCoords, endPts)
- coordinates += GlyphCoordinates(delta) * scalar
-
- glyph = copy(glyfTable[self.name]) # Shallow copy
- width, lsb, height, tsb = _setCoordinates(glyph, coordinates, glyfTable)
- self.lsb = lsb
- self.tsb = tsb
- if glyphSet.hvarTable is None:
- # no HVAR: let's set metrics from the phantom points
- self.width = width
- self.height = height
- return glyph
-
-
-class _TTGlyphCFF(_TTGlyph):
- def draw(self, pen):
- """Draw the glyph onto ``pen``. See fontTools.pens.basePen for details
- how that works.
- """
- self.glyphSet.charStrings[self.name].draw(pen, self.glyphSet.blender)
-
-
-def _setCoordinates(glyph, coord, glyfTable):
- # Handle phantom points for (left, right, top, bottom) positions.
- assert len(coord) >= 4
- leftSideX = coord[-4][0]
- rightSideX = coord[-3][0]
- topSideY = coord[-2][1]
- bottomSideY = coord[-1][1]
-
- for _ in range(4):
- del coord[-1]
-
- if glyph.isComposite():
- assert len(coord) == len(glyph.components)
- glyph.components = [copy(comp) for comp in glyph.components] # Shallow copy
- for p, comp in zip(coord, glyph.components):
- if hasattr(comp, "x"):
- comp.x, comp.y = p
- elif glyph.isVarComposite():
- glyph.components = [copy(comp) for comp in glyph.components] # Shallow copy
- for comp in glyph.components:
- coord = comp.setCoordinates(coord)
- assert not coord
- elif glyph.numberOfContours == 0:
- assert len(coord) == 0
- else:
- assert len(coord) == len(glyph.coordinates)
- glyph.coordinates = coord
-
- glyph.recalcBounds(glyfTable)
-
- horizontalAdvanceWidth = otRound(rightSideX - leftSideX)
- verticalAdvanceWidth = otRound(topSideY - bottomSideY)
- leftSideBearing = otRound(glyph.xMin - leftSideX)
- topSideBearing = otRound(topSideY - glyph.yMax)
- return (
- horizontalAdvanceWidth,
- leftSideBearing,
- verticalAdvanceWidth,
- topSideBearing,
- )
diff --git a/spaces/cihyFjudo/fairness-paper-search/Chhota Bheem and the throne of Bali telugu movie in hindi download Watch the adventure of Bheem and his friends.md b/spaces/cihyFjudo/fairness-paper-search/Chhota Bheem and the throne of Bali telugu movie in hindi download Watch the adventure of Bheem and his friends.md
deleted file mode 100644
index 6af4919e4ae57a9859d57625378ad8a186464e06..0000000000000000000000000000000000000000
--- a/spaces/cihyFjudo/fairness-paper-search/Chhota Bheem and the throne of Bali telugu movie in hindi download Watch the adventure of Bheem and his friends.md
+++ /dev/null
@@ -1,6 +0,0 @@
-
Chhota Bheem and the throne of Bali telugu movie in hindi download
-
- aaccfb2cb3
-
-
-
diff --git a/spaces/cihyFjudo/fairness-paper-search/The-Art-Of-Speculation-Philip-L-Carret-1930-Revised-Editionpdf.md b/spaces/cihyFjudo/fairness-paper-search/The-Art-Of-Speculation-Philip-L-Carret-1930-Revised-Editionpdf.md
deleted file mode 100644
index 431e4eec81d54a5ab168b5e9d75dc47d77f8648f..0000000000000000000000000000000000000000
--- a/spaces/cihyFjudo/fairness-paper-search/The-Art-Of-Speculation-Philip-L-Carret-1930-Revised-Editionpdf.md
+++ /dev/null
@@ -1,94 +0,0 @@
-## The Art Of Speculation Philip L Carret 1930 Revised Editionpdf
-
-
-
-
-
- 
-
-
-
-
-
-**Download — [https://www.google.com/url?q=https%3A%2F%2Furluso.com%2F2txlgJ&sa=D&sntz=1&usg=AOvVaw3t1HOH5DBwDd5vbDIOTMBq](https://www.google.com/url?q=https%3A%2F%2Furluso.com%2F2txlgJ&sa=D&sntz=1&usg=AOvVaw3t1HOH5DBwDd5vbDIOTMBq)**
-
-
-
-
-
-
-
-
-
-
-
-
-
-# The Art of Speculation: A Classic Book on Investing by Philip L. Carret
-
-
-
-Have you ever wondered what it takes to be a successful speculator in the stock market? Do you want to learn from one of the pioneers of mutual funds and a role model for Warren Buffett? If so, you might be interested in reading *The Art of Speculation* by Philip L. Carret.
-
-
-
-*The Art of Speculation* is a book that was first published in 1927 and revised in 1930 by Philip L. Carret, a famed investor and founder of The Pioneer Fund, one of the first mutual funds in the United States. Carret was a former Barron's reporter and WWI aviator who launched the fund in 1928 after managing money for his friends and family. He ran the fund for 55 years, during which an investment of $10,000 became $8 million. Warren Buffett said of him that he had "the best long term investment record of anyone I know".
-
-
-
-In this book, Carret shares his insights and wisdom on the art and science of speculation, which he defines as "the purchase or sale of securities or commodities in expectation of profiting by fluctuations in their prices". He covers topics such as the machinery of markets, the vehicles of speculation, market movements, forecasting the major swings, reading financial statements, analyzing different types of stocks, trading in unlisted securities, options and arbitrage, and when speculation becomes investment. He also provides examples and anecdotes from his own experience and from other famous speculators such as Jesse Livermore, Bernard Baruch, and J.P. Morgan.
-
-
-
-*The Art of Speculation* is a classic book on investing that has stood the test of time and is still relevant today. It is not a book for beginners, but rather for those who have some knowledge and experience in the stock market and want to improve their skills and judgment. It is also a book that requires careful study and reflection, as Carret does not offer any easy formulas or rules, but rather principles and guidelines that need to be applied with discretion and common sense.
-
-
-
-If you are interested in reading *The Art of Speculation* by Philip L. Carret, you can find a free pdf version online at [archive.org](https://archive.org/details/artofspeculation0000carr_x7y1). You can also buy a paperback edition at [Google Books](https://books.google.com/books/about/The_Art_Of_Speculation.html?id=ANFvCwAAQBAJ) or [Google Books](https://books.google.com/books/about/The_Art_of_Speculation.html?id=OfWnbHN3aQ8C).
-
-
-
-In this article, we will take a closer look at the life and career of Philip L. Carret, the author of *The Art of Speculation* and one of the most influential investors of the 20th century.
-
-
-
-## Early Life and Education
-
-
-
-Philip L. Carret was born on November 29, 1896 in Lynn, Massachusetts. He was interested in chemistry and mathematics from an early age and graduated from Harvard College in 1917 with a Bachelor of Science in Chemistry. He then attended Harvard Business School but did not complete his degree due to his enlistment in the U.S. Army Air Service during World War I. He served as a pilot and instructor until 1919.
-
-
-
-## Career as a Journalist and Investor
-
-
-
-After the war, Carret worked as a reporter for Barron's, a financial weekly magazine. He covered topics such as banking, insurance, railroads, and utilities. He also developed his own style of investing based on value principles, which he introduced in a series of articles in 1927. He advocated buying stocks that were undervalued by the market and holding them for the long term, regardless of short-term fluctuations. He also emphasized the importance of diversification, margin of safety, and fundamental analysis.
-
-
-
-In 1928, Carret founded one of the first mutual funds in the United States, The Pioneer Fund (then: Fidelity Mutual Trust). He started the fund with $25,000 of his own money and $500,000 from his friends and family. He managed the fund for 55 years, achieving an average annual return of 13% and turning an initial investment of $10,000 into $8 million. He was one of the few investors who survived the Great Depression and profited from it by buying stocks at bargain prices. He also invested in companies that were pioneers in their fields, such as IBM, Xerox, Polaroid, and Boeing.
-
-
-
-In 1963, Carret founded Carret Asset Management, a family office and investment advisory firm for institutional clients and high net worth families. He continued to manage money until his death in 1998 at age 101. He was known for his discipline, patience, humility, and curiosity. He was an avid reader and traveler who visited more than 100 countries and witnessed several solar eclipses.
-
-
-
-## Legacy and Influence
-
-
-
-Philip L. Carret was widely respected and admired by his peers and successors in the investment world. He was a role model for Warren Buffett, who said that Carret had "the best long term investment record of anyone I know". He was also praised by Benjamin Graham, John Templeton, Peter Lynch, and John Bogle. He received many honors and awards for his achievements, such as the Financial Analysts Federation Award for Outstanding Achievement in 1976 and the Harvard Business School Alumni Achievement Award in 1984.
-
-
-
-*The Art of Speculation* is one of Carret's most enduring contributions to the field of investing. It is a book that combines theory and practice, history and philosophy, wisdom and wit. It is a book that teaches not only how to invest but also how to think. It is a book that has inspired generations of investors who seek to master the art and science of speculation.
-
- 1b8d091108
-
-
-
-
-
diff --git a/spaces/cloudtheboi/Lofi4All/.pythonlibs/lib/python3.10/site-packages/fontTools/ttLib/tables/__init__.py b/spaces/cloudtheboi/Lofi4All/.pythonlibs/lib/python3.10/site-packages/fontTools/ttLib/tables/__init__.py
deleted file mode 100644
index f4cba26bf6ecaf18e96a62db69f70078498451e3..0000000000000000000000000000000000000000
--- a/spaces/cloudtheboi/Lofi4All/.pythonlibs/lib/python3.10/site-packages/fontTools/ttLib/tables/__init__.py
+++ /dev/null
@@ -1,96 +0,0 @@
-# DON'T EDIT! This file is generated by MetaTools/buildTableList.py.
-def _moduleFinderHint():
- """Dummy function to let modulefinder know what tables may be
- dynamically imported. Generated by MetaTools/buildTableList.py.
-
- >>> _moduleFinderHint()
- """
- from . import B_A_S_E_
- from . import C_B_D_T_
- from . import C_B_L_C_
- from . import C_F_F_
- from . import C_F_F__2
- from . import C_O_L_R_
- from . import C_P_A_L_
- from . import D_S_I_G_
- from . import D__e_b_g
- from . import E_B_D_T_
- from . import E_B_L_C_
- from . import F_F_T_M_
- from . import F__e_a_t
- from . import G_D_E_F_
- from . import G_M_A_P_
- from . import G_P_K_G_
- from . import G_P_O_S_
- from . import G_S_U_B_
- from . import G__l_a_t
- from . import G__l_o_c
- from . import H_V_A_R_
- from . import J_S_T_F_
- from . import L_T_S_H_
- from . import M_A_T_H_
- from . import M_E_T_A_
- from . import M_V_A_R_
- from . import O_S_2f_2
- from . import S_I_N_G_
- from . import S_T_A_T_
- from . import S_V_G_
- from . import S__i_l_f
- from . import S__i_l_l
- from . import T_S_I_B_
- from . import T_S_I_C_
- from . import T_S_I_D_
- from . import T_S_I_J_
- from . import T_S_I_P_
- from . import T_S_I_S_
- from . import T_S_I_V_
- from . import T_S_I__0
- from . import T_S_I__1
- from . import T_S_I__2
- from . import T_S_I__3
- from . import T_S_I__5
- from . import T_T_F_A_
- from . import V_D_M_X_
- from . import V_O_R_G_
- from . import V_V_A_R_
- from . import _a_n_k_r
- from . import _a_v_a_r
- from . import _b_s_l_n
- from . import _c_i_d_g
- from . import _c_m_a_p
- from . import _c_v_a_r
- from . import _c_v_t
- from . import _f_e_a_t
- from . import _f_p_g_m
- from . import _f_v_a_r
- from . import _g_a_s_p
- from . import _g_c_i_d
- from . import _g_l_y_f
- from . import _g_v_a_r
- from . import _h_d_m_x
- from . import _h_e_a_d
- from . import _h_h_e_a
- from . import _h_m_t_x
- from . import _k_e_r_n
- from . import _l_c_a_r
- from . import _l_o_c_a
- from . import _l_t_a_g
- from . import _m_a_x_p
- from . import _m_e_t_a
- from . import _m_o_r_t
- from . import _m_o_r_x
- from . import _n_a_m_e
- from . import _o_p_b_d
- from . import _p_o_s_t
- from . import _p_r_e_p
- from . import _p_r_o_p
- from . import _s_b_i_x
- from . import _t_r_a_k
- from . import _v_h_e_a
- from . import _v_m_t_x
-
-
-if __name__ == "__main__":
- import doctest, sys
-
- sys.exit(doctest.testmod().failed)
diff --git a/spaces/cncn102/bingo1/src/components/theme-toggle.tsx b/spaces/cncn102/bingo1/src/components/theme-toggle.tsx
deleted file mode 100644
index 67d3f1a2c163ccbeb52c40a7e42f107190237154..0000000000000000000000000000000000000000
--- a/spaces/cncn102/bingo1/src/components/theme-toggle.tsx
+++ /dev/null
@@ -1,31 +0,0 @@
-'use client'
-
-import * as React from 'react'
-import { useTheme } from 'next-themes'
-
-import { Button } from '@/components/ui/button'
-import { IconMoon, IconSun } from '@/components/ui/icons'
-
-export function ThemeToggle() {
- const { setTheme, theme } = useTheme()
- const [_, startTransition] = React.useTransition()
-
- return (
-
- )
-}
diff --git a/spaces/codedog-ai/codedog-demo/codedog_demo/callbacks.py b/spaces/codedog-ai/codedog-demo/codedog_demo/callbacks.py
deleted file mode 100644
index 552aca72c2766c79b2beb4dfbfc9c2760527cbde..0000000000000000000000000000000000000000
--- a/spaces/codedog-ai/codedog-demo/codedog_demo/callbacks.py
+++ /dev/null
@@ -1,61 +0,0 @@
-import time
-import traceback
-from functools import lru_cache
-from os import environ as env
-from os import listdir
-from typing import List
-
-import requests
-
-from codedog_demo.github_utils import parse_github_pr_url
-
-codedog_api = env.get("CODEDOG_API_URL", "")
-github_token = env.get("GITHUB_TOKEN", "")
-
-sample_names = []
-sample_contents = []
-for file in listdir("samples"):
- sample_names.append(file.replace("@", "/"))
- with open("samples/" + file, "r") as f:
- sample_contents.append(f.read())
-
-
-def request_pr_review(url: str):
- try:
- repo, pr_number = parse_github_pr_url(url)
- if not repo or not pr_number:
- return "Invalid URL. Accept format is: https://www.github.com/{owner}/{repository}/pull/{pr_number}", ""
-
- result = _request_pr_review(repo, pr_number, ttl_hash=get_ttl_hash())
-
- except Exception:
- traceback.print_exc()
- return "Something went wrong. Please try again later.", ""
- return result, result
-
-
-@lru_cache(maxsize=100)
-def _request_pr_review(repo: str, pr_number: int, ttl_hash=None):
- response = requests.post(
- codedog_api, json={"repository": repo, "pull_request_number": pr_number, "token": github_token}
- )
- result = response.text
- if len(result) < 100:
- if result == "stream timeout":
- raise ValueError("Timeout")
- print(f"Error result: {result}")
- raise ValueError()
- return result
-
-
-def get_ttl_hash(seconds=120):
- """Return the same value withing `seconds` time period"""
- return round(time.time() / seconds)
-
-
-def get_sample_choices() -> List[str]:
- return sample_names
-
-
-def show_sample(idx: int) -> str:
- return sample_contents[idx]
diff --git a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aac_ac3_parser.c b/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aac_ac3_parser.c
deleted file mode 100644
index 9ab979632dc98637c19eb7302203bcc85ed5ff1a..0000000000000000000000000000000000000000
--- a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aac_ac3_parser.c
+++ /dev/null
@@ -1,168 +0,0 @@
-/*
- * Common AAC and AC-3 parser
- * Copyright (c) 2003 Fabrice Bellard
- * Copyright (c) 2003 Michael Niedermayer
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include "config_components.h"
-
-#include "libavutil/channel_layout.h"
-#include "libavutil/common.h"
-#include "parser.h"
-#include "aac_ac3_parser.h"
-#include "ac3_parser_internal.h"
-#include "adts_header.h"
-
-int ff_aac_ac3_parse(AVCodecParserContext *s1,
- AVCodecContext *avctx,
- const uint8_t **poutbuf, int *poutbuf_size,
- const uint8_t *buf, int buf_size)
-{
- AACAC3ParseContext *s = s1->priv_data;
- ParseContext *pc = &s->pc;
- int len, i;
- int new_frame_start;
- int got_frame = 0;
-
- if (s1->flags & PARSER_FLAG_COMPLETE_FRAMES) {
- i = buf_size;
- got_frame = 1;
- } else {
-get_next:
- i=END_NOT_FOUND;
- if(s->remaining_size <= buf_size){
- if(s->remaining_size && !s->need_next_header){
- i= s->remaining_size;
- s->remaining_size = 0;
- }else{ //we need a header first
- len=0;
- for(i=s->remaining_size; istate = (s->state<<8) + buf[i];
- if((len=s->sync(s->state, &s->need_next_header, &new_frame_start)))
- break;
- }
- if(len<=0){
- i=END_NOT_FOUND;
- }else{
- got_frame = 1;
- s->state=0;
- i-= s->header_size -1;
- s->remaining_size = len;
- if(!new_frame_start || pc->index+i<=0){
- s->remaining_size += i;
- goto get_next;
- }
- else if (i < 0) {
- s->remaining_size += i;
- }
- }
- }
- }
-
- if(ff_combine_frame(pc, i, &buf, &buf_size)<0){
- s->remaining_size -= FFMIN(s->remaining_size, buf_size);
- *poutbuf = NULL;
- *poutbuf_size = 0;
- return buf_size;
- }
- }
-
- *poutbuf = buf;
- *poutbuf_size = buf_size;
-
- if (got_frame) {
- int bit_rate;
-
- /* Due to backwards compatible HE-AAC the sample rate, channel count,
- and total number of samples found in an AAC ADTS header are not
- reliable. Bit rate is still accurate because the total frame
- duration in seconds is still correct (as is the number of bits in
- the frame). */
- if (avctx->codec_id != AV_CODEC_ID_AAC) {
- AC3HeaderInfo hdr, *phrd = &hdr;
- int offset = ff_ac3_find_syncword(buf, buf_size);
-
- if (offset < 0)
- return i;
-
- buf += offset;
- buf_size -= offset;
- while (buf_size > 0) {
- int ret = avpriv_ac3_parse_header(&phrd, buf, buf_size);
-
- if (ret < 0 || hdr.frame_size > buf_size)
- return i;
-
- if (buf_size > hdr.frame_size) {
- buf += hdr.frame_size;
- buf_size -= hdr.frame_size;
- continue;
- }
- /* Check for false positives since the syncword is not enough.
- See section 6.1.2 of A/52. */
- if (av_crc(s->crc_ctx, 0, buf + 2, hdr.frame_size - 2))
- return i;
- break;
- }
-
- avctx->sample_rate = hdr.sample_rate;
-
- if (hdr.bitstream_id > 10)
- avctx->codec_id = AV_CODEC_ID_EAC3;
-
- if (!CONFIG_EAC3_DECODER || avctx->codec_id != AV_CODEC_ID_EAC3) {
- av_channel_layout_uninit(&avctx->ch_layout);
- if (hdr.channel_layout) {
- av_channel_layout_from_mask(&avctx->ch_layout, hdr.channel_layout);
- } else {
- avctx->ch_layout.order = AV_CHANNEL_ORDER_UNSPEC;
- avctx->ch_layout.nb_channels = hdr.channels;
- }
-#if FF_API_OLD_CHANNEL_LAYOUT
-FF_DISABLE_DEPRECATION_WARNINGS
- avctx->channels = avctx->ch_layout.nb_channels;
- avctx->channel_layout = hdr.channel_layout;
-FF_ENABLE_DEPRECATION_WARNINGS
-#endif
- }
- s1->duration = hdr.num_blocks * 256;
- avctx->audio_service_type = hdr.bitstream_mode;
- if (hdr.bitstream_mode == 0x7 && hdr.channels > 1)
- avctx->audio_service_type = AV_AUDIO_SERVICE_TYPE_KARAOKE;
- bit_rate = hdr.bit_rate;
- } else {
- AACADTSHeaderInfo hdr, *phrd = &hdr;
- int ret = avpriv_adts_header_parse(&phrd, buf, buf_size);
-
- if (ret < 0)
- return i;
-
- bit_rate = hdr.bit_rate;
- }
-
- /* Calculate the average bit rate */
- s->frame_number++;
- if (!CONFIG_EAC3_DECODER || avctx->codec_id != AV_CODEC_ID_EAC3) {
- avctx->bit_rate +=
- (bit_rate - avctx->bit_rate) / s->frame_number;
- }
- }
-
- return i;
-}
diff --git a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aacenc.h b/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aacenc.h
deleted file mode 100644
index b030c652aec88ac94a7f1d17186cb685a603ce36..0000000000000000000000000000000000000000
--- a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/aacenc.h
+++ /dev/null
@@ -1,162 +0,0 @@
-/*
- * AAC encoder
- * Copyright (C) 2008 Konstantin Shishkov
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#ifndef AVCODEC_AACENC_H
-#define AVCODEC_AACENC_H
-
-#include "libavutil/channel_layout.h"
-#include "libavutil/float_dsp.h"
-#include "libavutil/mem_internal.h"
-
-#include "avcodec.h"
-#include "put_bits.h"
-
-#include "aac.h"
-#include "audio_frame_queue.h"
-#include "psymodel.h"
-
-#include "lpc.h"
-
-typedef enum AACCoder {
- AAC_CODER_ANMR = 0,
- AAC_CODER_TWOLOOP,
- AAC_CODER_FAST,
-
- AAC_CODER_NB,
-}AACCoder;
-
-typedef struct AACEncOptions {
- int coder;
- int pns;
- int tns;
- int ltp;
- int pce;
- int pred;
- int mid_side;
- int intensity_stereo;
-} AACEncOptions;
-
-struct AACEncContext;
-
-typedef struct AACCoefficientsEncoder {
- void (*search_for_quantizers)(AVCodecContext *avctx, struct AACEncContext *s,
- SingleChannelElement *sce, const float lambda);
- void (*encode_window_bands_info)(struct AACEncContext *s, SingleChannelElement *sce,
- int win, int group_len, const float lambda);
- void (*quantize_and_encode_band)(struct AACEncContext *s, PutBitContext *pb, const float *in, float *out, int size,
- int scale_idx, int cb, const float lambda, int rtz);
- void (*encode_tns_info)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*encode_ltp_info)(struct AACEncContext *s, SingleChannelElement *sce, int common_window);
- void (*encode_main_pred)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*adjust_common_pred)(struct AACEncContext *s, ChannelElement *cpe);
- void (*adjust_common_ltp)(struct AACEncContext *s, ChannelElement *cpe);
- void (*apply_main_pred)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*apply_tns_filt)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*update_ltp)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*ltp_insert_new_frame)(struct AACEncContext *s);
- void (*set_special_band_scalefactors)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*search_for_pns)(struct AACEncContext *s, AVCodecContext *avctx, SingleChannelElement *sce);
- void (*mark_pns)(struct AACEncContext *s, AVCodecContext *avctx, SingleChannelElement *sce);
- void (*search_for_tns)(struct AACEncContext *s, SingleChannelElement *sce);
- void (*search_for_ltp)(struct AACEncContext *s, SingleChannelElement *sce, int common_window);
- void (*search_for_ms)(struct AACEncContext *s, ChannelElement *cpe);
- void (*search_for_is)(struct AACEncContext *s, AVCodecContext *avctx, ChannelElement *cpe);
- void (*search_for_pred)(struct AACEncContext *s, SingleChannelElement *sce);
-} AACCoefficientsEncoder;
-
-extern const AACCoefficientsEncoder ff_aac_coders[];
-
-typedef struct AACQuantizeBandCostCacheEntry {
- float rd;
- float energy;
- int bits;
- char cb;
- char rtz;
- uint16_t generation;
-} AACQuantizeBandCostCacheEntry;
-
-typedef struct AACPCEInfo {
- AVChannelLayout layout;
- int num_ele[4]; ///< front, side, back, lfe
- int pairing[3][8]; ///< front, side, back
- int index[4][8]; ///< front, side, back, lfe
- uint8_t config_map[16]; ///< configs the encoder's channel specific settings
- uint8_t reorder_map[16]; ///< maps channels from lavc to aac order
-} AACPCEInfo;
-
-/**
- * AAC encoder context
- */
-typedef struct AACEncContext {
- AVClass *av_class;
- AACEncOptions options; ///< encoding options
- PutBitContext pb;
- AVTXContext *mdct1024; ///< long (1024 samples) frame transform context
- av_tx_fn mdct1024_fn;
- AVTXContext *mdct128; ///< short (128 samples) frame transform context
- av_tx_fn mdct128_fn;
- AVFloatDSPContext *fdsp;
- AACPCEInfo pce; ///< PCE data, if needed
- float *planar_samples[16]; ///< saved preprocessed input
-
- int profile; ///< copied from avctx
- int needs_pce; ///< flag for non-standard layout
- LPCContext lpc; ///< used by TNS
- int samplerate_index; ///< MPEG-4 samplerate index
- int channels; ///< channel count
- const uint8_t *reorder_map; ///< lavc to aac reorder map
- const uint8_t *chan_map; ///< channel configuration map
-
- ChannelElement *cpe; ///< channel elements
- FFPsyContext psy;
- struct FFPsyPreprocessContext* psypp;
- const AACCoefficientsEncoder *coder;
- int cur_channel; ///< current channel for coder context
- int random_state;
- float lambda;
- int last_frame_pb_count; ///< number of bits for the previous frame
- float lambda_sum; ///< sum(lambda), for Qvg reporting
- int lambda_count; ///< count(lambda), for Qvg reporting
- enum RawDataBlockType cur_type; ///< channel group type cur_channel belongs to
-
- AudioFrameQueue afq;
- DECLARE_ALIGNED(16, int, qcoefs)[96]; ///< quantized coefficients
- DECLARE_ALIGNED(32, float, scoefs)[1024]; ///< scaled coefficients
-
- uint16_t quantize_band_cost_cache_generation;
- AACQuantizeBandCostCacheEntry quantize_band_cost_cache[256][128]; ///< memoization area for quantize_band_cost
-
- void (*abs_pow34)(float *out, const float *in, const int size);
- void (*quant_bands)(int *out, const float *in, const float *scaled,
- int size, int is_signed, int maxval, const float Q34,
- const float rounding);
-
- struct {
- float *samples;
- } buffer;
-} AACEncContext;
-
-void ff_aac_dsp_init_x86(AACEncContext *s);
-void ff_aac_coder_init_mips(AACEncContext *c);
-void ff_quantize_band_cost_cache_init(struct AACEncContext *s);
-
-
-#endif /* AVCODEC_AACENC_H */
diff --git a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/arm/h264chroma_init_arm.c b/spaces/colakin/video-generater/public/ffmpeg/libavcodec/arm/h264chroma_init_arm.c
deleted file mode 100644
index 5c7d5231865580968b130e3413f17eac9ba15db3..0000000000000000000000000000000000000000
--- a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/arm/h264chroma_init_arm.c
+++ /dev/null
@@ -1,57 +0,0 @@
-/*
- * ARM NEON optimised H.264 chroma functions
- * Copyright (c) 2008 Mans Rullgard
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-#include
-
-#include "libavutil/attributes.h"
-#include "libavutil/cpu.h"
-#include "libavutil/arm/cpu.h"
-#include "libavcodec/h264chroma.h"
-
-void ff_put_h264_chroma_mc8_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-void ff_put_h264_chroma_mc4_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-void ff_put_h264_chroma_mc2_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-
-void ff_avg_h264_chroma_mc8_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-void ff_avg_h264_chroma_mc4_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-void ff_avg_h264_chroma_mc2_neon(uint8_t *dst, const uint8_t *src, ptrdiff_t stride,
- int h, int x, int y);
-
-av_cold void ff_h264chroma_init_arm(H264ChromaContext *c, int bit_depth)
-{
- const int high_bit_depth = bit_depth > 8;
- int cpu_flags = av_get_cpu_flags();
-
- if (have_neon(cpu_flags) && !high_bit_depth) {
- c->put_h264_chroma_pixels_tab[0] = ff_put_h264_chroma_mc8_neon;
- c->put_h264_chroma_pixels_tab[1] = ff_put_h264_chroma_mc4_neon;
- c->put_h264_chroma_pixels_tab[2] = ff_put_h264_chroma_mc2_neon;
-
- c->avg_h264_chroma_pixels_tab[0] = ff_avg_h264_chroma_mc8_neon;
- c->avg_h264_chroma_pixels_tab[1] = ff_avg_h264_chroma_mc4_neon;
- c->avg_h264_chroma_pixels_tab[2] = ff_avg_h264_chroma_mc2_neon;
- }
-}
diff --git a/spaces/conciomith/RetinaFace_FaceDetector_Extractor/RetinaFace.py b/spaces/conciomith/RetinaFace_FaceDetector_Extractor/RetinaFace.py
deleted file mode 100644
index b3d0719a8cce330350d97fae7f5b5978bb6a64a5..0000000000000000000000000000000000000000
--- a/spaces/conciomith/RetinaFace_FaceDetector_Extractor/RetinaFace.py
+++ /dev/null
@@ -1,214 +0,0 @@
-import os
-os.environ['TF_CPP_MIN_LOG_LEVEL'] = '3'
-
-#---------------------------
-
-import numpy as np
-import tensorflow as tf
-import cv2
-
-import retinaface_model
-import preprocess
-import postprocess
-
-#---------------------------
-
-import tensorflow as tf
-tf_version = int(tf.__version__.split(".")[0])
-
-if tf_version == 2:
- import logging
- tf.get_logger().setLevel(logging.ERROR)
-
-#---------------------------
-
-def build_model():
-
- global model #singleton design pattern
-
- if not "model" in globals():
-
- model = tf.function(
- retinaface_model.build_model(),
- input_signature=(tf.TensorSpec(shape=[None, None, None, 3], dtype=np.float32),)
- )
-
- return model
-
-def get_image(img_path):
- if type(img_path) == str: # Load from file path
- if not os.path.isfile(img_path):
- raise ValueError("Input image file path (", img_path, ") does not exist.")
- img = cv2.imread(img_path)
-
- elif isinstance(img_path, np.ndarray): # Use given NumPy array
- img = img_path.copy()
-
- else:
- raise ValueError("Invalid image input. Only file paths or a NumPy array accepted.")
-
- # Validate image shape
- if len(img.shape) != 3 or np.prod(img.shape) == 0:
- raise ValueError("Input image needs to have 3 channels at must not be empty.")
-
- return img
-
-def detect_faces(img_path, threshold=0.9, model = None, allow_upscaling = True):
- """
- TODO: add function doc here
- """
-
- img = get_image(img_path)
-
- #---------------------------
-
- if model is None:
- model = build_model()
-
- #---------------------------
-
- nms_threshold = 0.4; decay4=0.5
-
- _feat_stride_fpn = [32, 16, 8]
-
- _anchors_fpn = {
- 'stride32': np.array([[-248., -248., 263., 263.], [-120., -120., 135., 135.]], dtype=np.float32),
- 'stride16': np.array([[-56., -56., 71., 71.], [-24., -24., 39., 39.]], dtype=np.float32),
- 'stride8': np.array([[-8., -8., 23., 23.], [ 0., 0., 15., 15.]], dtype=np.float32)
- }
-
- _num_anchors = {'stride32': 2, 'stride16': 2, 'stride8': 2}
-
- #---------------------------
-
- proposals_list = []
- scores_list = []
- landmarks_list = []
- im_tensor, im_info, im_scale = preprocess.preprocess_image(img, allow_upscaling)
- net_out = model(im_tensor)
- net_out = [elt.numpy() for elt in net_out]
- sym_idx = 0
-
- for _idx, s in enumerate(_feat_stride_fpn):
- _key = 'stride%s'%s
- scores = net_out[sym_idx]
- scores = scores[:, :, :, _num_anchors['stride%s'%s]:]
-
- bbox_deltas = net_out[sym_idx + 1]
- height, width = bbox_deltas.shape[1], bbox_deltas.shape[2]
-
- A = _num_anchors['stride%s'%s]
- K = height * width
- anchors_fpn = _anchors_fpn['stride%s'%s]
- anchors = postprocess.anchors_plane(height, width, s, anchors_fpn)
- anchors = anchors.reshape((K * A, 4))
- scores = scores.reshape((-1, 1))
-
- bbox_stds = [1.0, 1.0, 1.0, 1.0]
- bbox_deltas = bbox_deltas
- bbox_pred_len = bbox_deltas.shape[3]//A
- bbox_deltas = bbox_deltas.reshape((-1, bbox_pred_len))
- bbox_deltas[:, 0::4] = bbox_deltas[:,0::4] * bbox_stds[0]
- bbox_deltas[:, 1::4] = bbox_deltas[:,1::4] * bbox_stds[1]
- bbox_deltas[:, 2::4] = bbox_deltas[:,2::4] * bbox_stds[2]
- bbox_deltas[:, 3::4] = bbox_deltas[:,3::4] * bbox_stds[3]
- proposals = postprocess.bbox_pred(anchors, bbox_deltas)
-
- proposals = postprocess.clip_boxes(proposals, im_info[:2])
-
- if s==4 and decay4<1.0:
- scores *= decay4
-
- scores_ravel = scores.ravel()
- order = np.where(scores_ravel>=threshold)[0]
- proposals = proposals[order, :]
- scores = scores[order]
-
- proposals[:, 0:4] /= im_scale
- proposals_list.append(proposals)
- scores_list.append(scores)
-
- landmark_deltas = net_out[sym_idx + 2]
- landmark_pred_len = landmark_deltas.shape[3]//A
- landmark_deltas = landmark_deltas.reshape((-1, 5, landmark_pred_len//5))
- landmarks = postprocess.landmark_pred(anchors, landmark_deltas)
- landmarks = landmarks[order, :]
-
- landmarks[:, :, 0:2] /= im_scale
- landmarks_list.append(landmarks)
- sym_idx += 3
-
- proposals = np.vstack(proposals_list)
- if proposals.shape[0]==0:
- landmarks = np.zeros( (0,5,2) )
- return np.zeros( (0,5) ), landmarks
- scores = np.vstack(scores_list)
- scores_ravel = scores.ravel()
- order = scores_ravel.argsort()[::-1]
-
- proposals = proposals[order, :]
- scores = scores[order]
- landmarks = np.vstack(landmarks_list)
- landmarks = landmarks[order].astype(np.float32, copy=False)
-
- pre_det = np.hstack((proposals[:,0:4], scores)).astype(np.float32, copy=False)
-
- #nms = cpu_nms_wrapper(nms_threshold)
- #keep = nms(pre_det)
- keep = postprocess.cpu_nms(pre_det, nms_threshold)
-
- det = np.hstack( (pre_det, proposals[:,4:]) )
- det = det[keep, :]
- landmarks = landmarks[keep]
-
- resp = {}
- for idx, face in enumerate(det):
-
- label = 'face_'+str(idx+1)
- resp[label] = {}
- resp[label]["score"] = face[4]
-
- resp[label]["facial_area"] = list(face[0:4].astype(int))
-
- resp[label]["landmarks"] = {}
- resp[label]["landmarks"]["right_eye"] = list(landmarks[idx][0])
- resp[label]["landmarks"]["left_eye"] = list(landmarks[idx][1])
- resp[label]["landmarks"]["nose"] = list(landmarks[idx][2])
- resp[label]["landmarks"]["mouth_right"] = list(landmarks[idx][3])
- resp[label]["landmarks"]["mouth_left"] = list(landmarks[idx][4])
-
- return resp
-
-def extract_faces(img_path, threshold=0.9, model = None, align = True, allow_upscaling = True):
-
- resp = []
-
- #---------------------------
-
- img = get_image(img_path)
-
- #---------------------------
-
- obj = detect_faces(img_path = img, threshold = threshold, model = model, allow_upscaling = allow_upscaling)
-
- if type(obj) == dict:
- for key in obj:
- identity = obj[key]
-
- facial_area = identity["facial_area"]
- facial_img = img[facial_area[1]: facial_area[3], facial_area[0]: facial_area[2]]
-
- if align == True:
- landmarks = identity["landmarks"]
- left_eye = landmarks["left_eye"]
- right_eye = landmarks["right_eye"]
- nose = landmarks["nose"]
- mouth_right = landmarks["mouth_right"]
- mouth_left = landmarks["mouth_left"]
-
- facial_img = postprocess.alignment_procedure(facial_img, right_eye, left_eye, nose)
-
- resp.append(facial_img[:, :, ::-1])
- #elif type(obj) == tuple:
-
- return resp
diff --git a/spaces/congsaPfin/Manga-OCR/logs/FIFA 18 V10 APK - Play with Your Favorite Teams and Players in the Latest Version 2023.md b/spaces/congsaPfin/Manga-OCR/logs/FIFA 18 V10 APK - Play with Your Favorite Teams and Players in the Latest Version 2023.md
deleted file mode 100644
index c286cddd0c5dd27d235c288ae2e83f5d53ca7b49..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/FIFA 18 V10 APK - Play with Your Favorite Teams and Players in the Latest Version 2023.md
+++ /dev/null
@@ -1,104 +0,0 @@
-
-
FIFA 18 APK + OBB Download: How to Enjoy the World's Game on Your Android Device
-
If you are a fan of football, you probably have heard of FIFA, the most popular and realistic football simulation game series in the world. And if you own an Android device, you might be wondering how you can play the latest installment of this series, FIFA 18, on your mobile phone or tablet.
Well, wonder no more, because in this article, we will show you how you can download and install FIFA 18 APK + OBB files on your Android device, and enjoy the amazing features and gameplay of this game wherever you go.
-
FIFA 18 is a football simulation game developed by EA Sports and released in September 2017 for various platforms, including Windows, PlayStation, Xbox, Nintendo Switch, and Android. It is the 25th edition of the FIFA series, and it features Cristiano Ronaldo as the cover star.
-
FIFA 18 has received positive reviews from critics and players alike, who praised its improved graphics, animations, gameplay, modes, and content. It is one of the best-selling games of all time, with over 24 million copies sold worldwide by the end of 2018.
-
But what makes FIFA 18 so special and fun to play? Let's take a look at some of its features and gameplay.
-
FIFA 18 Features and Gameplay
-
FIFA 18 is not just a simple update of its predecessor, FIFA 17. It introduces several new features and improvements that make it stand out from other football games. Here are some of them:
-
Real Player Motion Technology
-
This is a new animation system that uses pose trajectory matching on every frame to deliver the most responsive and fluid gameplay ever. It captures the real movements and motions of top players like Ronaldo, Messi, Neymar, and more, making them look and feel like their real-life counterparts.
-
With Real Player Motion Technology, you can experience realistic player acceleration, deceleration, turns, sprints, dribbles, shots, passes, tackles, and more. You can also see how players react to different situations on the pitch, such as fatigue, pressure, collisions, injuries, etc.
-
Player Personality
-
This is another feature that adds more realism and authenticity to the game. It reflects how players behave and move on the pitch according to their unique characteristics and styles. For example, Ronaldo will run and dribble with his signature sprint and chop, Messi will weave through defenders with his agile and nimble movements, Neymar will show off his flair and skills with his tricks and flicks, etc.
-
fifa 18 v10 apk obb latest version 2023
-fifa 18 android apk obb offline download
-fifa 18 mobile apk obb mod download
-fifa 18 apk obb highly compressed download
-fifa 18 apk obb data file download
-fifa 18 apk obb free download for android
-fifa 18 ultimate team apk obb download
-fifa 18 world cup apk obb download
-fifa 18 apk obb full version download
-fifa 18 apk obb update patch download
-fifa 18 apk obb rexdl download
-fifa 18 apk obb revdl download
-fifa 18 apk obb mega download
-fifa 18 apk obb mediafire download
-fifa 18 apk obb google drive download
-fifa 18 apk obb direct link download
-fifa 18 apk obb no verification download
-fifa 18 apk obb unlocked download
-fifa 18 apk obb original download
-fifa 18 apk obb cracked download
-fifa 18 mod apk obb unlimited money download
-fifa 18 mod apk obb real faces download
-fifa 18 mod apk obb new kits download
-fifa 18 mod apk obb latest transfers download
-fifa 18 mod apk obb offline mode download
-how to install fifa 18 apk obb on android
-how to play fifa 18 apk obb online
-how to fix fifa 18 apk obb error
-how to update fifa 18 apk obb manually
-how to extract fifa 18 apk obb zip file
-best site to download fifa 18 apk obb
-best settings for fifa 18 apk obb
-best graphics for fifa 18 apk obb
-best players in fifa 18 apk obb
-best teams in fifa 18 apk obb
-tips and tricks for fifa 18 apk obb
-cheats and hacks for fifa 18 apk obb
-reviews and ratings for fifa 18 apk obb
-features and gameplay of fifa 18 apk obb
-requirements and compatibility of fifa 18 apk obb
-
Player Personality also affects how players interact with each other on the pitch, creating more realistic team chemistry and dynamics. For example, players will celebrate together after scoring a goal, console each other after missing a chance, argue with the referee or the opponents, etc.
-
Enhanced Dribbling and Crossing
-
FIFA 18 gives you more options and control on the ball, allowing you to create more chances and score more goals. You can use different types of dribbles, such as close control, speed dribble, skill dribble, etc., to beat defenders and create space. You can also use different types of crosses, such as early cross, driven cross, lofted cross, etc., to deliver accurate and dangerous balls into the box.
-
With Enhanced Dribbling and Crossing, you can unleash your creativity and style on the pitch, and enjoy the thrill of scoring spectacular goals.
-
The Journey: Hunter Returns
-
This is the second season of the story mode that debuted in FIFA 17. It follows the career of Alex Hunter, a young and talented footballer who dreams of becoming a star. You can control his actions and decisions on and off the pitch, affecting his relationships, reputation, and performance.
-
In FIFA 18, you can experience new challenges and opportunities as Alex Hunter moves to different clubs and leagues around the world. You can also customize his appearance, clothing, hairstyle, tattoos, etc., to suit your preferences. You can also meet and interact with famous players and managers, such as Ronaldo, Griezmann, Mourinho, etc.
-
The Journey: Hunter Returns is a captivating and immersive mode that lets you live the life of a footballer.
-
Other Modes and Content
-
FIFA 18 offers a variety of modes and content to suit your preferences. You can play online or offline, solo or with friends, casual or competitive. Here are some of the modes and content you can enjoy in FIFA 18:
- - Ultimate Team: This is the most popular mode in FIFA 18. It allows you to build your own dream team from scratch using players from different clubs and leagues. You can earn coins by playing matches or trading players on the market. You can also use FIFA Points to buy packs that contain random players or items. You can compete in various tournaments and seasons online or offline, and earn rewards such as coins, packs, players, etc. - Career Mode: This is the mode where you can manage your own club or play as a single player. You can choose from hundreds of clubs from different leagues around the world. You can scout for new players, negotiate contracts, set tactics, train your squad, etc. You can also play as a single player and improve your skills and attributes by completing objectives and tasks. You can also request for transfers or loans to other clubs. - Kick Off: This is the mode where you can play a quick match against the AI or another player. You can choose from any club or national team in the game. You can also customize the match settings such as difficulty level, half length, weather, stadium, etc. You can also play in different modes such as Classic, Women's, World Cup, etc. - Online Seasons: This is the mode where you can play online matches against other players of similar skill level. You can choose from any club or national team in the game. You can play in 10 divisions, each with 10 matches. You can earn points by winning or drawing matches, and move up or down the divisions based on your performance. You can also earn coins and trophies by completing seasons. - Online Friendlies: This is the mode where you can play online matches against your friends. You can invite your friends to join your session, and choose from any club or national team in the game. You can also customize the match settings such as difficulty level, half length, weather, stadium, etc. You can also track your stats and results against your friends. - Skill Games: This is the mode where you can practice and improve your skills in various aspects of the game. You can choose from different categories such as shooting, passing, dribbling, defending, etc. You can also play in different levels of difficulty from beginner to advanced. You can earn coins and badges by completing skill games. - Customize: This is the mode where you can customize various aspects of the game to suit your preferences. You can edit players, teams, leagues, stadiums, balls, kits, etc. You can also create your own custom tournaments and leagues. You can also download and apply updates and patches for the game.
FIFA 18 System Requirements and Compatibility
-
If you want to play FIFA 18 on your Android device, you need to make sure that your device meets the minimum and recommended specifications for the game. Here are the system requirements and compatibility for FIFA 18:
-
Minimum Specifications
- - Android version: 4.4 KitKat or higher - CPU: Quad-core 1.4 GHz or higher - RAM: 1 GB or higher - Storage: 2 GB or higher - Internet connection: Required for online features
Recommended Specifications
- - Android version: 6.0 Marshmallow or higher - CPU: Octa-core 2.0 GHz or higher - RAM: 2 GB or higher - Storage: 4 GB or higher - Internet connection: Required for online features
Supported Devices
-
FIFA 18 is compatible with most Android devices that meet the minimum specifications. However, some devices may run the game better than others, depending on their hardware and software configurations. Here are some of the supported devices for FIFA 18:
- | Device | Model | | --- | --- | | Samsung | Galaxy S7, S8, S9, Note 8, Note 9, A5, A7, J5, J7 | | Huawei | P10, P20, Mate 10, Mate 20, Honor 8, Honor 9 | | LG | G5, G6, G7, V20, V30 | | Sony | Xperia XZ1, XZ2 | | Motorola | Moto G5S Plus, Moto Z2 Play | | Xiaomi | Mi 6X, Mi A1, Mi A2, Redmi Note 5, Redmi Note 6 | | OnePlus | OnePlus 5, OnePlus 6 | | Google | Pixel 2, Pixel 3 | | Nokia | Nokia 6, Nokia 7 Plus | | Asus | Zenfone 4, Zenfone 5 |
This is not a complete list of supported devices, and you may be able to run FIFA 18 on other devices as well. However, if your device is not on this list, you may experience some issues or errors while playing the game.
-
Installation Guide and Tips
-
To download and install FIFA 18 APK + OBB files on your Android device, you need to follow these steps:
- - Step 1: Download the FIFA 18 APK + OBB files from a trusted source. You can use the link below to download them directly from our website. The files are safe and virus-free, and they have been tested and verified by us. - Step 2: Enable the installation of apps from unknown sources on your device. To do this, go to Settings > Security > Unknown Sources and toggle it on. This will allow you to install apps that are not from the Google Play Store. - Step 3: Locate the downloaded FIFA 18 APK + OBB files on your device. You can use a file manager app to find them in your Downloads folder or any other folder where you saved them. - Step 4: Install the FIFA 18 APK file by tapping on it and following the instructions on the screen. Do not open the app yet after the installation is complete. - Step 5: Extract the FIFA 18 OBB file using a zip extractor app. You will get a folder named com.ea.gp.fifaworld. Copy this folder and paste it in your Android > OBB folder. If you don't have an OBB folder, you can create one yourself. - Step 6: Launch the FIFA 18 app from your app drawer or home screen. You may need to verify your identity with an email or phone number. You may also need to download some additional data for the game to run smoothly. - Step 7: Enjoy playing FIFA 18 on your Android device.
Here are some tips to help you play FIFA 18 better on your Android device:
- - Tip 1: Make sure you have enough storage space and battery life on your device before playing the game. FIFA 18 is a large and demanding game that requires at least 2 GB of storage space and a lot of battery power to run properly. - Tip 2: Adjust the graphics settings and controls according to your device's performance and your preference. You can access these settings from the main menu of the game. You can lower the graphics quality and resolution to improve the framerate and reduce lag. You can also change the control scheme and sensitivity to suit your playstyle. - Tip 3: Connect to a stable and fast internet connection when playing online modes or features. FIFA 18 requires an internet connection to access some of its modes and features, such as Ultimate Team, Online Seasons, Online Friendlies, etc. You can use Wi-Fi or mobile data, but make sure they are reliable and fast enough to avoid disconnections or delays. - Tip 4: Update the game regularly to get the latest features and fixes. FIFA 18 receives frequent updates from EA Sports that add new content, improve gameplay, fix bugs, etc. You can update the game automatically or manually from the Google Play Store or from our website.
FIFA 18 Review and Rating
-
FIFA 18 is one of the best football games ever made, and it has received rave reviews from critics and players alike. Here are some of the reviews and ratings for FIFA 18:
-
Critic Reviews
-
FIFA 18 has an average score of 84 out of 100 on Metacritic, based on the reviews of 41 critics. Here are some of the excerpts from the reviews:
- - "FIFA 18 is simply magnificent. Streets ahead of what came before, and continuing its dominance over its rivals, EA has done a superb job. With huge improvements across the board, this is the game FIFA fans have waited five years for." - Trusted Reviews - "FIFA 18 is a far better football game than its predecessor. I was rather fond of FIFA 17, but despite the engine overhaul it was still beholden to some of FIFA’s more long-standing issues. Animations taking too long to unfold and delaying your move; wrestling to control unresponsive players; a lack of individuality from player to player." - IGN - "FIFA 18 is the best FIFA game EA has ever made. It’s that simple. I cannot believe the huge leap the series has made in one year. This is streets ahead of FIFA 17, let alone any that came before that." - The Sun
User Reviews
-
FIFA 18 has an average score of 6.1 out of 10 on Metacritic, based on the reviews of 1,223 users. Here are some of the excerpts from the reviews:
- - "FIFA 18 is a great game with amazing graphics and gameplay. The Journey mode is very interesting and fun to play. The Ultimate Team mode is addictive and rewarding. The online modes are smooth and competitive. The best FIFA game ever!" - User Review - "FIFA 18 is a good game but not a great one. The graphics and animations are impressive but the gameplay is still flawed and inconsistent. The Journey mode is boring and repetitive. The Ultimate Team mode is pay-to-win and unfair. The online modes are laggy and frustrating. The same FIFA game every year!" - User Review - "FIFA 18 is a terrible game with awful graphics and gameplay. The Journey mode is a joke and a waste of time. The Ultimate Team mode is a scam and a rip-off. The online modes are broken and unplayable. The worst FIFA game ever!" - User Review
Pros and Cons
-
FIFA 18 has its strengths and weaknesses, like any other game. Here are some of the pros and cons of FIFA 18:
- | Pros | Cons | | --- | --- | | - Stunning graphics and animations | - High system requirements and compatibility issues | | - Realistic and fluid gameplay | - Flawed and inconsistent gameplay | | - Captivating and immersive story mode | - Boring and repetitive story mode | | - Variety of modes and content | - Pay-to-win and unfair mode | | - Online features and community | - Online issues and problems |
Conclusion
-
FIFA 18 is a football simulation game that offers a lot of features and gameplay options for fans of the sport. It has improved graphics, animations, gameplay, modes, and content compared to its previous versions. It also has a story mode that follows the career of Alex Hunter, a young footballer who wants to become a star.
-
However, FIFA 18 also has some drawbacks and limitations that may affect your enjoyment of the game. It has high system requirements and compatibility issues that may prevent you from running the game smoothly on your Android device. It also has flawed and inconsistent gameplay that may frustrate you at times. It also has pay-to-win and unfair modes that may discourage you from playing them.
-
Overall, FIFA 18 is a great game for football lovers, but it is not perfect. You may love it or hate it depending on your expectations and preferences.
-
If you want to try FIFA 18 on your Android device, you can download and install it using the link below. You can also check out our website for more games, apps, tips, tricks, guides, etc.
-
Thank you for reading this article, and we hope you have fun playing FIFA 18 on your Android device.
-
FAQs
-
Here are some of the frequently asked questions about FIFA 18:
-
Q1: Is FIFA 18 free to download and play on Android?
-
A1: Yes, FIFA 18 is free to download and play on Android devices. However, you may need to pay for some in-game items or features using real money or FIFA Points.
-
Q2: How much storage space do I need to install FIFA 18 on my Android device?
-
A2: You need at least 2 GB of storage space to install FIFA 18 on your Android device . However, you may need more space to download additional data or updates for the game.
-
Q3: How can I update FIFA 18 to the latest version on my Android device?
-
A3: You can update FIFA 18 to the latest version on your Android device by using the Google Play Store or our website. You can check for updates manually or enable automatic updates from the settings of the app. You can also download and install the latest APK + OBB files from our website.
-
Q4: How can I play online with other players in FIFA 18 on my Android device?
-
A4: You can play online with other players in FIFA 18 on your Android device by using an internet connection and logging in to your EA account. You can play online modes such as Ultimate Team, Online Seasons, Online Friendlies, etc. You can also join online communities and chat with other players.
-
Q5: How can I fix common issues or errors in FIFA 18 on my Android device?
-
A5: You can fix common issues or errors in FIFA 18 on your Android device by following these steps:
- - Step 1: Make sure your device meets the minimum and recommended specifications for the game. - Step 2: Make sure you have enough storage space and battery life on your device. - Step 3: Make sure you have a stable and fast internet connection. - Step 4: Update the game to the latest version. - Step 5: Clear the cache and data of the app. - Step 6: Restart your device and launch the app again. - Step 7: Contact EA support if the problem persists. 401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/FR Legends APK 3.3.1 - The Ultimate Drifting Game for Android Devices.md b/spaces/congsaPfin/Manga-OCR/logs/FR Legends APK 3.3.1 - The Ultimate Drifting Game for Android Devices.md
deleted file mode 100644
index b8f316b1896d313c863af2b9ca026b7550a22cc2..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/FR Legends APK 3.3.1 - The Ultimate Drifting Game for Android Devices.md
+++ /dev/null
@@ -1,95 +0,0 @@
-
-
FR Legends APK 3.3.1: The Ultimate Drifting Game for Android
-
If you are a fan of drifting and racing games, you might want to check out FR Legends APK 3.3.1, the latest version of the popular mobile game that lets you experience the spirit of drifting like never before.
FR Legends is a mobile game developed by Feng Li, a Chinese indie developer who is passionate about drifting and motorsports. The game was released in 2018 and has since gained a huge fan base around the world, especially among drift enthusiasts and car lovers.
-
FR Legends stands for "Front-engine, Rear-wheel-drive Legends", which refers to the type of cars that are used for drifting, such as Toyota AE86, Nissan Silvia, Mazda RX-7, and more. The game allows you to choose from a variety of drift cars, customize them to your liking, and compete with other players online or offline in various game modes.
-
Features of FR Legends
-
FR Legends is not just a simple racing game, it is a game that simulates the art and culture of drifting in a realistic and fun way. Here are some of the features that make FR Legends stand out from other drifting games:
-
Customizable cars
-
One of the most appealing aspects of FR Legends is that you can customize your own drift car to suit your style and preference. You can change the color, body kit, wheels, tires, suspension, engine, exhaust, and more. You can also add stickers, decals, and accessories to make your car look unique and cool.
FR Legends uses a realistic physics engine that makes the drifting experience more authentic and challenging. You have to master the throttle, brake, steering, and handbrake to control your car's angle and speed while drifting. You also have to deal with tire wear, smoke, damage, and collisions.
-
Online multiplayer
-
FR Legends lets you compete with other players from around the world in online multiplayer mode. You can join or create a room with up to six players and race against each other in tandem or solo mode. You can also chat with other players and make friends or rivals.
-
Various game modes
-
FR Legends offers different game modes to suit your mood and skill level. You can play in career mode, where you have to complete various missions and challenges to earn money and reputation. You can also play in free mode, where you can practice your drifting skills without any pressure or rules. You can also play in arcade mode, where you can enjoy some fun and casual drifting games.
-
How to download and install FR Legends APK 3.3.1
-
If you want to play FR Legends on your Android device, you have to download and install the APK file from a trusted source. Here are the requirements and steps to do so:
-
Requirements
-
-
Your Android device must have at least Android 5.0 or higher.
-
Your Android device must have at least 100 MB of free storage space.
-
You must enable the installation of apps from unknown sources on your Android device.
-
-
Steps
-
-
Download the FR Legends APK 3.3.1 file from this link.
-
Locate the downloaded file on your Android device and tap on it to start the installation process.
-
Follow the instructions on the screen to complete the installation.
-
Launch the game and enjoy!
-Pros and cons of FR Legends APK 3.3.1
-
FR Legends APK 3.3.1 is not a perfect game, it has its pros and cons. Here are some of them:
-
Pros
-
-
It is free to download and play.
-
It has amazing graphics and sound effects.
-
It has a large and active community of players and fans.
-
It has frequent updates and new features.
-
It is fun and addictive to play.
-
-
Cons
-
-
It may not be compatible with some devices or regions.
-
It may have some bugs or glitches.
-
It may require a stable internet connection for online mode.
-
It may be too hard or frustrating for some players.
-
It may have some ads or in-app purchases.
-
-
Conclusion
-
FR Legends APK 3.3.1 is a game that will make you feel the thrill and excitement of drifting in a realistic and fun way. You can customize your own drift car, compete with other players online or offline, and enjoy various game modes. If you are a fan of drifting and racing games, you should definitely give FR Legends a try. You will not regret it!
-
FAQs
-
Here are some frequently asked questions about FR Legends APK 3.3.1:
-
-
What is the difference between FR Legends APK and FR Legends MOD APK? The FR Legends APK is the original version of the game that you can download from the official source. The FR Legends MOD APK is a modified version of the game that may have some extra features or cheats, such as unlimited money, unlocked cars, etc. However, the FR Legends MOD APK may not be safe or legal to use, so we do not recommend it.
-
How can I get more money in FR Legends? You can get more money in FR Legends by completing missions and challenges in career mode, winning races in online mode, watching ads, or buying in-app purchases.
-
How can I play FR Legends on PC? You can play FR Legends on PC by using an Android emulator, such as BlueStacks, NoxPlayer, or LDPlayer. You have to download and install the emulator on your PC, then download and install the FR Legends APK file on the emulator, and then launch the game from the emulator.
-
How can I contact the developer of FR Legends? You can contact the developer of FR Legends by sending an email to frlegends@outlook.com, or by following their social media accounts on Facebook, Twitter, Instagram, or YouTube.
-
Is FR Legends safe to download and play? Yes, FR Legends is safe to download and play, as long as you download it from a trusted source, such as this link. You should also scan the APK file with an antivirus software before installing it on your device.
- 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/FateGrand Order Mod APK with Menu Damage and Easy Win Features.md b/spaces/congsaPfin/Manga-OCR/logs/FateGrand Order Mod APK with Menu Damage and Easy Win Features.md
deleted file mode 100644
index db06e2700695af0ca9846dbaaa0357e5ceb59667..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/FateGrand Order Mod APK with Menu Damage and Easy Win Features.md
+++ /dev/null
@@ -1,118 +0,0 @@
-
-
Fate/Grand Order APK Mods: How to Find and Install Them from Reddit
-
Fate/Grand Order is a mobile game that has taken the world by storm. Based on the popular Fate franchise by Type-Moon, the game lets you summon and command historical, mythical, and fictional heroes known as Servants to fight against enemies that threaten the human history. With stunning graphics, engaging story, and diverse gameplay, Fate/Grand Order has attracted millions of fans across the globe.
-
However, not everyone is satisfied with the official version of the game. Some players want to have more control over their gameplay, such as increasing their damage, skipping battles, or getting unlimited resources. That's why some players resort to using APK mods, which are modified versions of the game that alter some of its features or functions.
But where can you find and install Fate/Grand Order APK mods? One of the most popular sources is Reddit, a social media platform where users can share and discuss various topics. In this article, we will show you how to find and install Fate/Grand Order APK mods from Reddit, as well as the pros and cons of using them.
-
What are APK Mods and Why Do Some Players Use Them?
-
An APK mod is a modified version of an Android application package (APK), which is the file format used to distribute and install applications on Android devices. By modifying the APK file, hackers or modders can change some aspects of the game, such as adding new features, removing restrictions, or altering the game data.
-
Some players use APK mods for various reasons, such as:
-
-
To gain an advantage in the game, such as increasing their damage, unlocking all Servants, or getting unlimited resources.
-
To bypass some limitations or difficulties in the game, such as skipping battles, avoiding ads, or accessing region-locked content.
-
To experience new or different aspects of the game, such as changing the graphics, adding new modes, or customizing their Servants.
-
-
However, using APK mods also comes with some risks and drawbacks, which we will discuss later in this article.
-
How to Find and Install Fate/Grand Order APK Mods from Reddit?
-
Reddit is one of the most popular sources of Fate/Grand Order APK mods. There are several subreddits (communities) dedicated to sharing and discussing Fate/Grand Order APK mods, such as r/grandordermods, r/FateGOmodding, or r/FateGOHacks. These subreddits often have links to download sites or guides on how to install the mods.
-
To find and install Fate/Grand Order APK mods from Reddit, you need to follow these steps:
-
fate grand order hack mod apk download
-fate grand order mod menu apk android
-fate grand order damage multiplier mod apk
-fate grand order easy win mod apk
-fate grand order blackmod team mod apk
-fate grand order reddit apk mod guide
-fate grand order reddit apk mod discussion
-fate grand order reddit apk mod review
-fate grand order reddit apk mod tips
-fate grand order reddit apk mod news
-fate grand order reddit apk mod update
-fate grand order reddit apk mod support
-fate grand order reddit apk mod help
-fate grand order reddit apk mod feedback
-fate grand order reddit apk mod request
-fate grand order reddit apk mod link
-fate grand order reddit apk mod source
-fate grand order reddit apk mod safe
-fate grand order reddit apk mod legit
-fate grand order reddit apk mod working
-fate grand order reddit apk mod latest
-fate grand order reddit apk mod version
-fate grand order reddit apk mod free
-fate grand order reddit apk mod premium
-fate grand order reddit apk mod vip
-fate grand order reddit apk mod unlimited
-fate grand order reddit apk mod features
-fate grand order reddit apk mod benefits
-fate grand order reddit apk mod advantages
-fate grand order reddit apk mod disadvantages
-fate grand order reddit apk mod pros and cons
-fate grand order reddit apk mod comparison
-fate grand order reddit apk mod alternatives
-fate grand order reddit apk mod recommendations
-fate grand order reddit apk mod suggestions
-fate grand order reddit apk mod questions
-fate grand order reddit apk mod answers
-fate grand order reddit apk mod solutions
-fate grand order reddit apk mod problems
-fate grand order reddit apk mod issues
-fate grand order reddit apk mod bugs
-fate grand order reddit apk mod fixes
-fate grand order reddit apk mod patches
-fate grand order reddit apk mod cheats
-fate grand order reddit apk mod hacks
-fate grand order reddit apk mod tricks
-fate grand order reddit apk mod secrets
-fate grand order reddit apk mod tutorials
-fate grand order reddit apk mod how to
-
-
Browse through the subreddits that offer Fate/Grand Order APK mods and look for a mod that suits your preferences. Make sure to read the description, comments, and reviews of the mod before downloading it.
-
Download the mod from a reliable and safe source. Avoid clicking on suspicious links or ads that may contain malware or viruses. You may need to use a VPN or proxy service if the download site is blocked in your region.
-
Backup your original Fate/Grand Order APK file and data before installing the mod. You can use a file manager app or a backup app to do this. This way, you can restore your original game if something goes wrong with the mod.
-
Uninstall your original Fate/Grand Order app from your device. You can do this by going to Settings > Apps > Fate/Grand Order > Uninstall.
-
Install the modded Fate/Grand Order APK file on your device. You may need to enable Unknown Sources in your device settings to allow installation from third-party sources.
-
Launch the modded Fate/Grand Order app and enjoy your modified gameplay.
-
-
Note: Some mods may require additional steps or files to work properly. Make sure to follow the instructions provided by the modder carefully.
-
What are the Pros and Cons of Using Fate Grand Order APK Mods?
-
Using Fate/Grand Order APK mods can have some benefits and drawbacks, depending on your perspective and preferences. Here are some of the pros and cons of using Fate/Grand Order APK mods:
-
-
-
Pros
-
Cons
-
-
-
You can have more fun and freedom in your gameplay, such as using your favorite Servants, skipping boring battles, or getting more rewards.
-
You may lose the challenge and satisfaction of playing the game as intended, such as overcoming difficult enemies, earning your resources, or following the story.
-
-
-
You can access content that is not available in your region, such as Japanese-only Servants, events, or voice lines.
-
You may violate the terms of service of the game and risk getting banned or suspended from the game or losing your account data.
-
-
-
You can experience new or different features that are not in the official version of the game, such as improved graphics, custom modes, or fan-made Servants.
-
You may encounter bugs, errors, or compatibility issues that may affect your gameplay or damage your device.
-
-
-
Ultimately, the decision to use Fate/Grand Order APK mods is up to you. You should weigh the pros and cons carefully and decide whether you are willing to take the risks or not.
-
Conclusion: Should You Use Fate/Grand Order APK Mods or Not?
-
Fate/Grand Order is a great game that offers a lot of entertainment and enjoyment for its fans. However, some players may want to modify their gameplay by using APK mods, which are modified versions of the game that change some of its features or functions.
-
There are many sources of Fate/Grand Order APK mods, but one of the most popular ones is Reddit, a social media platform where users can share and discuss various topics. You can find and install Fate/Grand Order APK mods from Reddit by following some simple steps, but you should also be aware of the pros and cons of using them.
-
Using Fate/Grand Order APK mods can have some benefits, such as having more fun and freedom in your gameplay, accessing region-locked content, or experiencing new or different features. However, it can also have some drawbacks, such as losing the challenge and satisfaction of playing the game as intended, violating the terms of service of the game and risking getting banned or suspended, or encountering bugs, errors, or compatibility issues.
-
Therefore, you should use Fate/Grand Order APK mods at your own discretion and responsibility. You should also respect the game developers and other players who play the game legitimately. Remember that Fate/Grand Order is a game that is meant to be enjoyed by everyone.
-
FAQs: Some Common Questions and Answers about Fate/Grand Order APK Mods
-
Here are some common questions and answers about Fate/Grand Order APK mods that you may find helpful:
-
Q: Are Fate/Grand Order APK mods legal?
-
A: The legality of Fate/Grand Order APK mods may vary depending on your country or region. Generally speaking, modifying an application without the permission of the developer is considered illegal and may infringe on their intellectual property rights. However, some countries or regions may have more lenient laws or regulations regarding this matter. You should check your local laws before using Fate/Grand Order APK mods.
-
Q: Are Fate/Grand Order APK mods safe?
-
A: The safety of Fate/Grand Order APK mods may depend on the source and quality of the mod. Some mods may be safe and harmless, while others may contain malware or viruses that may harm your device or steal your personal information. You should always download Fate/Grand Order APK mods from reliable and trustworthy sources and scan them with an antivirus software before installing them. You should also backup your original game data before using any mod.
-
Q: How do I update my Fate/Grand Order APK mod?
-
A: The update process of Fate/Grand Order APK mod may vary depending on the type and version of the mod. Some mods may update automatically or have an update option within the app. Others may require you to download and install a new version of the mod manually. You should always check the modder's website or Reddit post for any updates or instructions regarding their mod.
-
Q: How do I uninstall my Fate/Grand Order APK mod?
-
A: The uninstall process of Fate/Grand Order APK mod may depend on how you installed it in the first place. If you installed it by replacing your original game app, you can simply uninstall it by going to Settings > Apps > Fate/Grand Order > Uninstall. If you installed it by using a parallel app or a clone app, you can uninstall it by going to the app settings and choosing the uninstall option. If you want to restore your original game app, you can reinstall it from the official source or from your backup file.
-
Q: Can I use Fate/Grand Order APK mods with other players or online features?
-
A: The compatibility of Fate/Grand Order APK mods with other players or online features may depend on the nature and extent of the mod. Some mods may work fine with other players or online features, while others may cause errors, crashes, or bans. You should always be careful and respectful when using Fate/Grand Order APK mods with other players or online features, as you may ruin their experience or violate the game rules.
-
-
This is the end of the article. I hope you found it helpful and informative. Thank you for reading.
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/How to Get Komodo Chess 14 for Free and Play Like a Grandmaster.md b/spaces/congsaPfin/Manga-OCR/logs/How to Get Komodo Chess 14 for Free and Play Like a Grandmaster.md
deleted file mode 100644
index 2eed664fba081f8e652663543fbaaf52cbb10298..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/How to Get Komodo Chess 14 for Free and Play Like a Grandmaster.md
+++ /dev/null
@@ -1,131 +0,0 @@
-
-
Komodo Chess 14 Free Download: How to Get the World Champion Chess Engine
-
If you are looking for a powerful and versatile chess engine that can help you improve your chess skills, you might want to check out Komodo Chess 14. This is the latest version of the world champion chess engine that has won several prestigious titles and awards. In this article, we will show you how to download Komodo Chess 14 for free, how to install and use it, and how to make the most of its features and benefits.
Komodo Chess 14 is a chess engine developed by GM Larry Kaufman and Mark Lefler, inspired by AlphaZero, the artificial intelligence program that defeated the best chess engines in the world. Komodo Chess 14 thinks like no other chess program, using a combination of brute force and human-like intuition to find the best moves in any position. It can play both standard chess and variants, such as Fischer Random, King of the Hill, Suicide, etc. It can also switch between different modes and personalities, such as MCTS (Monte Carlo Tree Search), Armageddon, Contempt, etc. Komodo Chess 14 is not only a strong opponent, but also a great teacher and analyzer, providing a Grandmaster evaluation of any position and suggesting improvements.
-
Features and benefits of Komodo Chess 14
-
Some of the features and benefits of Komodo Chess 14 are:
-
-
It is the three-time TCEC champion, the most prestigious online computer chess event, and has also won several CCT events and several World Championships.
-
It is a significant strength improvement over the previous version, about 12 elo in MCTS mode and 10 elo in standard mode.
-
It has a new feature called "Armageddon" mode, which tells Komodo that White (or Black) must win, draws are scored as losses for that color. This improves Komodo's performance as White by about 30 elo.
-
It has more levels, personalities, and auto-skill features added since Komodo 13.
-
It can play both standard chess and variants, such as Fischer Random, King of the Hill, Suicide, etc.
-
It can switch between different modes and personalities, such as MCTS (Monte Carlo Tree Search), Armageddon, Contempt, etc.
-
It provides a Grandmaster evaluation of any position and suggests improvements.
-
-
How to download Komodo Chess 14 for free
-
If you want to download Komodo Chess 14 for free, you have two options:
-
-
You can visit the official website of Komodo Chess and click on the "Free" tab. There you will find a link to download Komodo 13.01 for free. This is an older version of Komodo Chess, but still very strong and useful.
-
You can visit the website of ChessBase, one of the compatible GUIs for Komodo Chess. There you will find a link to download a free trial version of Komodo Chess 14. This is a limited version of Komodo Chess 14 that expires after one month. However, you can still use it to test its features and performance.
-
-
How to install and use Komodo Chess 14
How to install and use Komodo Chess 14
-
Once you have downloaded Komodo Chess 14, either for free or as a paid product, you need to install it and use it with a compatible GUI (Graphical User Interface). A GUI is a software that allows you to interact with the chess engine, such as setting up the board, playing games, analyzing positions, etc. Komodo Chess 14 does not come with its own GUI, so you need to use one of the following options:
-
System requirements and compatible GUIs
-
The system requirements for Komodo Chess 14 are:
-
-
A 64-bit operating system (Windows, Linux, or Mac OS)
-
A 64-bit processor (Intel or AMD)
-
At least 4 GB of RAM
-
At least 100 MB of free disk space
-
-
The compatible GUIs for Komodo Chess 14 are:
-
komodo chess 14.1 free download
-komodo dragon chess engine free download
-komodo 14 chess software free download
-how to download komodo chess 14 for free
-komodo chess 14 vs stockfish 14 free download
-komodo chess 14 world champion edition free download
-komodo chess 14 review and free download
-komodo chess 14 system requirements and free download
-komodo chess 14 features and benefits free download
-komodo chess 14 installation guide and free download
-komodo chess 14 mcts mode free download
-komodo chess 14 armageddon mode free download
-komodo chess 14 personalities and auto-skill free download
-komodo chess 14 compatible guis free download
-komodo chess 14 multi-core support free download
-komodo chess 14 evaluation developed by a grandmaster free download
-komodo chess 14 three-time tcec champion free download
-komodo chess 14 world computer chess champion free download
-komodo chess 14 world chess software champion free download
-komodo chess 14 world computer blitz champion free download
-komodo chess 14 inspired by alphazero free download
-komodo chess 14 redeveloped from the ground up free download
-komodo chess 14 improved king safety and time management free download
-komodo chess 14 best settings and options free download
-komodo chess 14 user manual and tutorials free download
-komodo chess 14 latest updates and patches free download
-komodo chess 14 customer reviews and testimonials free download
-komodo chess 14 discount code and coupon free download
-komodo chess 14 official site and support free download
-komodo chess 14 vs other chess engines free download
-best way to learn from komodo chess 14 free download
-how to play against komodo chess 14 online free download
-how to analyze your games with komodo chess 14 free download
-how to improve your rating with komodo chess 14 free download
-how to train your opening repertoire with komodo chess 14 free download
-how to master the endgame with komodo chess 14 free download
-how to use the dragon by komodo chess app free download
-how to get the android version of komodo chess 14 free download
-how to get the mac osx version of komodo chess 14 free download
-how to get the linux version of komodo chess 14 free download
-how to get the windows version of komodo chess 14 free download
-how to get the serial number for komodo chess 14 free download
-how to activate your license for komodo chess 14 free download
-how to upgrade from previous versions of komodo chess for free
-
-
ChessBase: This is the most popular and professional chess software in the world. It has a huge database of games, a powerful analysis tool, and many other features. You can buy Komodo Chess 14 as a standalone product or as part of a bundle with ChessBase. You can also download a free trial version of Komodo Chess 14 from ChessBase.
-
Fritz: This is another chess software from the same company as ChessBase. It has similar features but is more user-friendly and less expensive. You can buy Komodo Chess 14 as a standalone product or as part of a bundle with Fritz. You can also download a free trial version of Komodo Chess 14 from Fritz.
-
Arena: This is a free and open-source chess software that supports many chess engines, including Komodo Chess 14. It has a simple and intuitive interface and some basic features. You can download Arena for free from its official website.
-
-
Installation instructions and tips
-
The installation process for Komodo Chess 14 depends on the GUI you are using. Here are some general steps and tips:
-
-
Download the ZIP file of Komodo Chess 14 from the official website or from the GUI website.
-
Extract the ZIP file to a folder on your computer.
-
Open the GUI of your choice and go to the menu where you can add or manage chess engines.
-
Select the option to add a new chess engine and browse to the folder where you extracted Komodo Chess 14.
-
Select the executable file of Komodo Chess 14 (either komodo-14.exe or komodo-14-mcts.exe) and click OK.
-
The GUI will recognize Komodo Chess 14 and add it to the list of available engines.
-
You can now select Komodo Chess 14 as your opponent or as your analyzer in the GUI.
-
-
Some tips to optimize the performance of Komodo Chess 14 are:
-
-
Make sure you have enough RAM and CPU power for Komodo Chess 14 to run smoothly.
-
Adjust the settings of Komodo Chess 14 according to your preferences and needs. You can change the parameters such as hash size, threads, contempt, etc. in the engine options menu of the GUI.
-
Use a large opening book and endgame tablebase for better results. You can download them from various sources online or buy them from the GUI websites.
-
How to play and analyze with Komodo Chess 14
-
Once you have installed and configured Komodo Chess 14, you can start playing and analyzing with it in the GUI of your choice. Here are some common ways to use Komodo Chess 14:
-
-
Play a game against Komodo Chess 14: You can choose the level, time control, color, and variant of the game. You can also enable or disable hints, takebacks, and engine assistance. You can see the evaluation, best move, and principal variation of Komodo Chess 14 during the game. You can also save, load, or export the game for later review.
-
Analyze a position or a game with Komodo Chess 14: You can set up any position on the board or load a game from a file or a database. You can then activate Komodo Chess 14 as your analyzer and see its evaluation, best move, and principal variation. You can also see the depth, nodes, speed, and score of Komodo Chess 14. You can adjust the analysis parameters such as multipv, infinite mode, etc. You can also add comments, variations, and annotations to the position or the game.
-
Use Komodo Chess 14 as a training tool: You can use various features of the GUI to improve your chess skills with Komodo Chess 14. For example, you can use the blunder check feature to find your mistakes in a game and see how Komodo Chess 14 would have played instead. You can also use the tactical analysis feature to generate puzzles from your games and see how Komodo Chess 14 would have solved them. You can also use the opening trainer feature to learn and practice openings with Komodo Chess 14.
-
-
How to improve your chess skills with Komodo Chess 14
-
Komodo Chess 14 is not only a strong opponent, but also a great teacher and analyzer. It can help you improve your chess skills in various ways. Here are some tips on how to use Komodo Chess 14 for learning and improvement:
-
Learn from the Grandmaster evaluation
-
Komodo Chess 14 provides a Grandmaster evaluation of any position and suggests improvements. You can learn from this evaluation by understanding why Komodo Chess 14 prefers certain moves over others, what are the plans and ideas behind them, and what are the strengths and weaknesses of each side. You can also compare your moves with Komodo Chess 14's moves and see where you went wrong or right. You can also ask Komodo Chess 14 to explain its moves in natural language using the "Why?" feature of the GUI.
-
Explore different modes and personalities
-
Komodo Chess 14 can switch between different modes and personalities, such as MCTS (Monte Carlo Tree Search), Armageddon, Contempt, etc. You can explore these modes and personalities by playing against them or analyzing with them. You can see how Komodo Chess 14 changes its style and behavior depending on the mode or personality. You can also learn from the different perspectives and approaches that Komodo Chess 14 offers. For example, you can use MCTS mode to see how Komodo Chess 14 thinks like AlphaZero, or use Armageddon mode to see how Komodo Chess 14 plays aggressively when it has to win.
-
Challenge yourself with puzzles and games
-
Komodo Chess 14 can also provide you with puzzles and games that challenge your chess skills. You can use various features of the GUI to generate puzzles from your games or from a database of games. You can then try to solve them with or without Komodo Chess 14's help. You can also play games against Komodo Chess 14 at different levels, time controls, colors, and variants. You can then review your games with Komodo Chess 14's analysis and feedback. You can also use the auto-skill feature of Komodo Chess 14 to adjust its level according to your performance.
-
Conclusion
-
Komodo Chess 14 is a world champion chess engine that thinks like no other chess program. It is a powerful and versatile chess engine that can play both standard chess and variants, switch between different modes and personalities, and provide a Grandmaster evaluation of any position. It is not only a strong opponent, but also a great teacher and analyzer that can help you improve your chess skills.
-
If you want to download Komodo Chess 14 for free, you have two options: either download an older version of Komodo Chess from its official website or download a free trial version of Komodo Chess 14 from one of its compatible GUIs (ChessBase or Fritz). If you want to buy Komodo Chess 14 as a paid product, you can do so from one of its compatible GUIs (ChessBase or Fritz). You can also buy Komodo Chess 14 as part of a bundle with ChessBase or Fritz, which will give you access to many other features and benefits. We hope this article has helped you learn more about Komodo Chess 14 and how to download it for free. If you are interested in trying out this amazing chess engine, don't hesitate to download it and start playing and analyzing with it. You will be amazed by how much you can improve your chess skills with Komodo Chess 14. Here are some FAQs that might answer some of your questions:
FAQs
-
-
What is the difference between Komodo 14 and Komodo 14 MCTS?
-
Komodo 14 is the standard version of Komodo Chess 14, which uses a brute force approach to search for the best moves. Komodo 14 MCTS is a variant of Komodo Chess 14, which uses a Monte Carlo Tree Search approach to search for the best moves. MCTS is inspired by AlphaZero, the artificial intelligence program that defeated the best chess engines in the world. MCTS is more creative and human-like than brute force, but also less reliable and consistent.
-
How can I update Komodo Chess 14?
-
If you have bought Komodo Chess 14 as a paid product, you can update it for free whenever a new version is released. You can do this by visiting the official website of Komodo Chess or the GUI website where you bought it and downloading the latest version. You can then install it over the previous version or in a new folder.
-
How can I contact the developers of Komodo Chess 14?
-
If you have any questions, feedback, or suggestions for the developers of Komodo Chess 14, you can contact them by visiting their official website and filling out the contact form. You can also join their forum and interact with other users and developers.
-
How can I support the development of Komodo Chess 14?
-
If you want to support the development of Komodo Chess 14, you can do so by buying their products, donating to their PayPal account, or subscribing to their Patreon page. You can also spread the word about Komodo Chess 14 and share your experiences with it on social media and online platforms.
-
Where can I find more information and resources about Komodo Chess 14?
-
If you want to find more information and resources about Komodo Chess 14, you can visit their official website, their forum, their YouTube channel, their Facebook page, their Twitter account, or their blog. You can also read reviews, articles, and books about Komodo Chess 14 online or offline.
- 401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/How to Save The Dog from the Angry Bees in this Awesome APK Puzzle Game.md b/spaces/congsaPfin/Manga-OCR/logs/How to Save The Dog from the Angry Bees in this Awesome APK Puzzle Game.md
deleted file mode 100644
index 97fc43fd40acadab1888545c334c5f8489cfe845..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/How to Save The Dog from the Angry Bees in this Awesome APK Puzzle Game.md
+++ /dev/null
@@ -1,214 +0,0 @@
-
-
Save The Dog Bee APK: A Fun and Challenging Puzzle Game
-
If you are looking for a casual puzzle game that will test your brain and reflexes, you might want to try Save The Dog Bee APK. This is a game where you have to draw lines with your fingers to create walls that protect the dog from attacks by bees in the hive. You need to protect the dog with the painted wall for 10 seconds during the attack of the bees, hold on and you will win the game. Use your brain to save the doge.
In this article, I will tell you more about Save The Dog Bee APK, such as its features, how to play, how to download and install it on your Android device, and some tips and tricks to help you win the game. I will also answer some frequently asked questions about the game. Let's get started!
-
Features of Save The Dog Bee APK
-
Save The Dog Bee APK is a simple but addictive puzzle game that will keep you entertained for hours. Here are some of the features that make this game fun and interesting:
-
-
A variety of levels
-
The game has hundreds of levels with different difficulty levels and challenges. You will never get bored as you try to save the dog from different scenarios and obstacles.
-
save the dog bee apk download
-save the dog bee apk mod
-save the dog bee apk free
-save the dog bee apk latest version
-save the dog bee apk android
-save the dog bee apk offline
-save the dog bee apk hack
-save the dog bee apk unlimited money
-save the dog bee apk no ads
-save the dog bee apk full version
-save the dog bee apk game
-save the dog bee apk puzzle
-save the dog bee apk funspace
-save the dog bee apk review
-save the dog bee apk tips
-save the dog bee apk cheats
-save the dog bee apk guide
-save the dog bee apk walkthrough
-save the dog bee apk gameplay
-save the dog bee apk trailer
-save the dog bee apk update
-save the dog bee apk new levels
-save the dog bee apk online
-save the dog bee apk multiplayer
-save the dog bee apk challenges
-save the dog bee apk skins
-save the dog bee apk characters
-save the dog bee apk graphics
-save the dog bee apk sound
-save the dog bee apk music
-save the dog bee apk rating
-save the dog bee apk size
-save the dog bee apk requirements
-save the dog bee apk compatibility
-save the dog bee apk installation
-save the dog bee apk support
-save the dog bee apk feedback
-save the dog bee apk alternatives
-save the dog bee apk similar games
-save the dog bee apk genre
-save the dog bee apks for pc
-save the dog bee apks for ios
-save the dog bee apks for windows 10
-save the dog bee apks for macbook
-save the dog bee apks for chromebook
-download Save The Dog APK (Android Game) - Free Download - APKCombo[^1^]
-tải xuống Save The Dog Bee APK cho Android[^2^]
-
Easy and funny gameplay
-
The game is easy to play but hard to master. You just need to swipe the screen to create a wall to protect the dog. As long as you don't let go, you can always draw the line. You can let go after producing a satisfactory pattern. Wait for the bees in the hive to attack. Hold your wall for 10 seconds, so that the dog will not be attacked by bees. You will win the game.
-
Funny dog expressions
-
The game has cute and funny graphics and animations. You will love the dog's expressions as he reacts to your actions and the bees' attacks. He will smile, cry, wink, or make other funny faces depending on the situation.
-
Puzzle and interesting levels
-
The game is not only about drawing lines. You also need to use your brain and logic to find the best way to save the dog. Sometimes you need to use other objects or tools in the environment, such as balloons, fans, magnets, or bombs. You also need to avoid traps and hazards that can harm the dog or break your wall.
-
Various skins
-
The game allows you to customize your dog with different skins. You can choose from different breeds, colors, or costumes. You can also save other animals besides dogs, such as chickens or sheep.
-
-
How to Play Save The Dog Bee APK
-
Save The Dog Bee APK is easy to play but challenging to master. Here are some basic steps on how to play the game:
-
-
Download and install Save The Dog Bee APK on your Android device.
-
Open the game and choose a level.
-
Swipe the screen to create a wall to protect the dog from the bees.
-
Wait for the bees in the hive to attack.
-
Hold your wall for 10 seconds without letting go.
-
If the dog survives without being stung by bees, you win the level.
-
If the dog gets stung by bees or your wall breaks, you lose the level.
-
Try again until you win or move on to another level.
-
-
How to Download and Install Save The Dog Bee APK on Android
-
If you want to download and install Save The Dog Bee APK on your Android device, you can follow these simple steps:
-
-
Go to [this link](^1^) or [this link](^2^) on your browser.
-
Tap on Download APK
Wait for the download to finish.
-
Go to your file manager and locate the downloaded APK file.
-
Tap on the file and allow installation from unknown sources if prompted.
-
Wait for the installation to complete.
-
Open the game and enjoy!
-
-
Tips and Tricks to Win Save The Dog Bee APK
-
Save The Dog Bee APK is a fun and challenging puzzle game that requires you to think fast and act smart. Here are some tips and tricks that can help you win the game:
-
-
Plan ahead
-
Before you start drawing your wall, take a look at the level and see where the bees are coming from, what obstacles are in the way, and what tools or objects you can use. Try to anticipate the bees' movements and draw your wall accordingly. You can also use the pause button to think more carefully.
-
Use different shapes
-
You don't have to draw a straight line to create a wall. You can also use curves, circles, triangles, or other shapes to protect the dog. Sometimes, using different shapes can help you cover more area or create more stability for your wall.
-
Be creative
-
You can also use your wall to interact with other elements in the level, such as balloons, fans, magnets, or bombs. You can use your wall to pop balloons, deflect fans, attract magnets, or detonate bombs. These can help you create more space or clear more bees for your dog.
-
Be careful
-
You also need to be careful not to harm your dog or break your wall. Avoid drawing your wall too close to the dog or too far from the bees. Also, avoid drawing your wall over traps or hazards that can damage your wall or hurt your dog.
-
Have fun
-
The most important tip is to have fun while playing Save The Dog Bee APK. Don't get frustrated if you lose a level or make a mistake. Just try again and enjoy the game!
-
-
Conclusion
-
Save The Dog Bee APK is a casual puzzle game that will test your brain and reflexes as you try to save the dog from attacks by bees in the hive. You need to draw lines with your fingers to create walls that protect the dog for 10 seconds. The game has hundreds of levels with different difficulty levels and challenges. You also need to use your logic and creativity to find the best way to save the dog. The game has cute and funny graphics and animations that will make you smile. You can also customize your dog with different skins or save other animals besides dogs.
-
If you want to download and install Save The Dog Bee APK on your Android device, you can follow the simple steps I mentioned above. You can also use some of the tips and tricks I shared with you to help you win the game. I hope you enjoyed this article and found it helpful. If you have any questions or feedback about Save The Dog Bee APK, feel free to leave a comment below.
-
Frequently Asked Questions
-
Here are some of the frequently asked questions about Save The Dog Bee APK:
-
-
Is Save The Dog Bee APK safe to download and install?
-
Yes, Save The Dog Bee APK is safe to download and install on your Android device. It does not contain any viruses, malware, or spyware that can harm your device or compromise your privacy. However, you should always download it from a trusted source like [this link] or [this link] and scan it with an antivirus app before installing it.
-
Is Save The Dog Bee APK free to play?
-
Yes, Save The Dog Bee APK is free to play on your Android device. You don't need to pay any money to download, install, or play the game. However, the game may contain some ads that can be removed by purchasing an ad-free version of the game.
-
How can I get more skins for my dog?
-
You can get more skins for your dog by completing levels and earning coins. You can use these coins to buy different skins from the shop. You can also watch ads or share the game with your friends to get more coins.
-
How can I save other animals besides dogs?
-
You can save other animals besides dogs by unlocking them from the shop. You need to have enough coins to buy them. Some of the animals you can save are chickens, sheep, pigs, cows, and cats.
-
What are the benefits of playing Save The Dog Bee APK?
-
Save The Dog Bee APK is not only a fun and entertaining game, but also a beneficial one. Playing this game can help you improve your cognitive skills, such as memory, attention, concentration, problem-solving, and creativity. It can also help you relieve stress, relax, and have a good mood.
-
-
Outline of the Article
-
Here is the outline of the article I wrote:
-
-
-
Heading
-
Subheading
-
Content
-
-
-
H1
-
Save The Dog Bee APK: A Fun and Challenging Puzzle Game
-
An introduction to the game and its main features.
-
-
-
H2
-
Features of Save The Dog Bee APK
-
A list of the features that make the game fun and interesting.
-
-
-
H3
-
A variety of levels
-
A description of the different levels and challenges in the game.
-
-
-
H3
-
Easy and funny gameplay
-
A description of how to play the game and its mechanics.
-
-
-
H3
-
Funny dog expressions
-
A description of the graphics and animations in the game.
-
-
-
H3
-
Puzzle and interesting levels
-
A description of how the game requires logic and creativity to save the dog.
-
-
-
H3
-
Various skins
-
A description of how to customize the dog with different skins or save other animals.
-
-
-
H2
-
How to Play Save The Dog Bee APK
-
A step-by-step guide on how to play the game.
-
-
-
H2
-
How to Download and Install Save The Dog Bee APK on Android
-
A step-by-step guide on how to download and install the game on Android devices.
-
-
H2
-
Tips and Tricks to Win Save The Dog Bee APK
-
A list of tips and tricks that can help the player win the game.
-
-
-
H3
-
Plan ahead
-
A tip on how to anticipate the bees' movements and draw the wall accordingly.
-
-
-
H3
-
Use different shapes
-
A tip on how to use curves, circles, triangles, or other shapes to create a wall.
-
-
-
H3
-
Be creative
-
A tip on how to use the wall to interact with other elements in the level.
-
-
-
H3
-
Be careful
-
A tip on how to avoid harming the dog or breaking the wall.
-
-
-
H3
-
Have fun
-
A tip on how to enjoy the game and not get frustrated.
-
-
-
H2
-
Conclusion
-
A summary of the main points of the article and a call to action for the reader.
-
-
-
H2
-
Frequently Asked Questions
-
A list of FAQs about the game and their answers.
-
-
I hope you liked my article and found it useful.
197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Incredibox APK iOS Make Music with a Merry Crew of Beatboxers.md b/spaces/congsaPfin/Manga-OCR/logs/Incredibox APK iOS Make Music with a Merry Crew of Beatboxers.md
deleted file mode 100644
index 36937cac6e38e504a4c38261bcfb4335a077dbc2..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Incredibox APK iOS Make Music with a Merry Crew of Beatboxers.md
+++ /dev/null
@@ -1,163 +0,0 @@
-
-
Incredibox Apk Ios: A Fun and Interactive Music App
-
Do you love music and want to create your own songs with a simple and intuitive app? If yes, then you should check out Incredibox apk ios, a music video game that lets you make music with a merry crew of beatboxers. In this article, we will tell you everything you need to know about Incredibox apk ios, including how to download and install it on your device, what are its features, what are the different versions available, and what are some of the best alternatives to it. Let's get started!
-
What is Incredibox and what can you do with it?
-
Incredibox is a music app that was created in 2009 by the French company So Far So Good (SFSG). It is a fun, interactive, and educational tool that allows you to create your own music with the help of a group of seven animated beatboxers. You can choose your musical style among nine impressive atmospheres and start to lay down, record, and share your mix. You can also find the right sound combos to unlock animated choruses that will enhance your tune. Incredibox is a great way to learn about rhythm and melody, as well as to express your creativity and musical talent.
Incredibox apk ios is the mobile version of Incredibox that is compatible with iOS devices such as iPhone, iPad, and iPod touch. You can download it from the App Store for $4.99. It requires iOS 15.0 or later and a Mac with Apple M1 chip or later. It has a 4.9-star rating on the App Store based on more than 30,000 reviews. It has also won several awards and appeared in various international media outlets such as BBC, Adobe, FWA, Gizmodo, Slate, Konbini, Softonic, Kotaku, Cosmopolitan, PocketGamer, AppAdvice, AppSpy, Vice, Ultralinx, and many others.
-
How to download and install Incredibox apk ios on your device?
-
Downloading and installing Incredibox apk ios on your device is very easy. Just follow these simple steps:
-
-
Go to the App Store on your device and search for "Incredibox".
-
Tap on the app icon and then tap on "Get" or "Buy" to purchase it.
-
Enter your Apple ID password or use Touch ID or Face ID to confirm your purchase.
-
Wait for the app to download and install on your device.
-
Once the app is installed, tap on "Open" or find it on your home screen.
-
Enjoy making music with Incredibox!
-
-
What are the main features of Incredibox apk ios?
-
Incredibox apk ios has many features that make it a fun and interactive music app. Here are some of them:
-
-
You can choose from nine different musical styles: Alpha (old school beatbox), Little Miss (R&B), Sunrise (pop), The Love (romantic), Brazil (samba), Alive (electro), Jeevan (Bollywood), Dystopia (cyberpunk), and Wekiddy (kids).
-
You can drag and drop icons onto the avatars to make them sing and start to compose your own music. Each icon represents a different sound loop such as beats, effects, melodies, chorus, or voices.
-
You can record your mix and share it with your friends or the world via email, social media, or the Incredibox website. You can also download your mix as an MP3 file or a video file.
-
You can explore the Top 50 chart on the Incredibox website and discover the best mixes created by other users. You can also vote for your favorite mixes and leave comments.
-
You can play with the app offline and enjoy it anywhere and anytime.
-
-
What are the different versions of Incredibox apk ios and how do they differ?
-
Incredibox apk ios has nine different versions that correspond to the nine musical styles available. Each version has its own theme, graphics, sounds, and bonuses. Here is a table that summarizes the main differences between the versions:
-
-
-
Version
-
Release Date
-
Theme
-
Bonus
-
-
-
Alpha
-
2009
-
Old school beatbox
-
A rap battle between two beatboxers
-
-
-
Little Miss
-
2012
-
R&B
-
A love story between a girl and a beatboxer
-
-
-
Sunrise
-
2013
-
Pop
-
A flash mob dance in a park
-
-
-
The Love
-
2014
-
Romantic
-
A wedding ceremony with a choir of beatboxers
-
-
-
Brazil
-
2016
-
Samba
-
A carnival parade with dancers and musicians
-
-
-
Alive
-
2017
-
Electro
-
A futuristic concert with robots and lasers
-
-
-
Jeevan
-
2018
-
Bollywood
-
A musical scene with dancers and elephants
-
-
-
Dystopia
-
2020
-
Cyberpunk
-
-
Wekiddy
-
2021
-
Kids
-
A playful scene with toys and animals
-
-
-
What are some of the best alternatives to Incredibox apk ios for music lovers?
-
If you enjoy Incredibox apk ios, you might also like some of these other music apps that let you create, record, and share your own tunes:
-
-
GarageBand: This is a popular app that turns your device into a full-featured recording studio. You can play, record, and mix music with a variety of instruments, loops, and effects. You can also collaborate with other musicians and share your songs via iCloud or social media.
-
Music Maker Jam: This is a fun app that lets you create your own music in minutes. You can choose from thousands of studio-quality loops, beats, and samples to mix and match your own tracks. You can also apply effects, change the tempo, and adjust the volume. You can also join a global community of music makers and discover new genres and styles.
-
Beat Snap: This is an easy-to-use app that lets you make beats and music with your fingers. You can tap on the pads to play sounds, record your performance, and edit it later. You can also add effects, filters, and vocals to spice up your tracks. You can also explore and remix songs from other users or upload your own to the cloud.
-
-
Conclusion
-
Incredibox apk ios is a fun and interactive music app that lets you make music with a merry crew of beatboxers. You can choose from nine different musical styles, drag and drop icons onto the avatars to make them sing, unlock animated bonuses, record and share your mix, and explore the Top 50 chart. Incredibox apk ios is a great way to learn about rhythm and melody, as well as to express your creativity and musical talent. You can download it from the App Store for $4.99 and enjoy it offline anywhere and anytime.
-
If you are looking for more music apps to try out, you can also check out GarageBand, Music Maker Jam, or Beat Snap. They are some of the best alternatives to Incredibox apk ios that let you create, record, and share your own tunes with a variety of instruments, loops, and effects.
-
We hope you found this article helpful and informative. If you have any questions or feedback, please feel free to leave a comment below. Thank you for reading!
-
incredibox app for ios
-incredibox music game ios
-incredibox download for iphone
-incredibox beatbox app ios
-incredibox create your own music ios
-incredibox v8 dystopia ios
-incredibox v7 jeevan ios
-incredibox v6 alive ios
-incredibox v5 brazil ios
-incredibox the love ios
-incredibox sunrise ios
-incredibox little miss ios
-incredibox app store
-incredibox ipad app
-incredibox ipod touch app
-incredibox mac app
-incredibox app review
-incredibox app price
-incredibox app features
-incredibox app privacy
-incredibox app dark mode
-incredibox app mp3 file
-incredibox app mixlist
-incredibox app for kids
-incredibox app no ads
-how to get incredibox on ios
-how to play incredibox on ios
-how to record incredibox on ios
-how to share incredibox on ios
-how to download incredibox on ios
-is incredibox available on ios
-is incredibox free on ios
-is incredibox safe on ios
-is incredibox offline on ios
-is incredibox worth it on ios
-best music apps like incredibox for ios
-best beatbox apps like incredibox for ios
-best game apps like incredibox for ios
-best creative apps like incredibox for ios
-best educational apps like incredibox for ios
-learn music with incredibox on ios
-make beats with incredibox on ios
-have fun with incredibox on ios
-enjoy the full incredibox experience on ios
-discover the 9 musical atmospheres of incredibox on ios
-join the top 50 chart of incredibox on ios
-watch the animated choruses of incredibox on ios
-explore the futuristic world of incrediobox on ios
-celebrate life with the mystic rhythm of incrediobox on ios
-
FAQs
-
Here are some of the frequently asked questions about Incredibox apk ios:
-
-
Is Incredibox apk ios free?
-
No, Incredibox apk ios is not free. It costs $4.99 on the App Store. However, there is no in-app purchase or subscription required to use the app.
-
Is Incredibox apk ios safe?
-
Yes, Incredibox apk ios is safe to use. It does not contain any harmful or malicious content. It also does not collect or share any personal or sensitive information from the users.
-
Is Incredibox apk ios compatible with Android devices?
-
No, Incredibox apk ios is not compatible with Android devices. It is only available for iOS devices such as iPhone, iPad, and iPod touch. However, there is an Android version of Incredibox that you can download from Google Play for $3.99.
-
How do I update Incredibox apk ios?
-
To update Incredibox apk ios, you need to go to the App Store on your device and check for any available updates. If there is an update available, you need to tap on "Update" or "Install" to download and install it on your device.
-
How do I contact the developers of Incredibox apk ios?
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/My Talking Angela APK iOS Tips and Tricks to Make Your Pet Happy.md b/spaces/congsaPfin/Manga-OCR/logs/My Talking Angela APK iOS Tips and Tricks to Make Your Pet Happy.md
deleted file mode 100644
index 0431c5f526b4d923f4e9564e30b31afff4b9063f..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/My Talking Angela APK iOS Tips and Tricks to Make Your Pet Happy.md
+++ /dev/null
@@ -1,94 +0,0 @@
-
-
My Talking Angela APK iOS: How to Download and Play the Fun Virtual Pet Game
-
If you are looking for a fun and cute virtual pet game for your iOS device, you might want to check out My Talking Angela. This game lets you adopt a stylish cat named Angela and take care of her as your own. You can dress her up, play with her, feed her, and watch her grow. In this article, we will show you how to download and play My Talking Angela APK iOS, as well as some tips and tricks to make the most out of your experience.
My Talking Angela is a virtual pet game developed by Outfit7, the same company behind the popular My Talking Tom series. It was released in 2014 and has since gained millions of fans around the world. The game is available for both Android and iOS devices, as well as Windows Phone and Amazon Kindle.
-
Features and activities of My Talking Angela
-
My Talking Angela is more than just a pet simulator. It is also a fashion star, a dancer, a singer, and a friend. You can enjoy various features and activities with Angela, such as:
-
-
Stylish makeup looks
-
Wonderful wardrobe choices
-
Super sweet activities
-
Special sticker albums
-
Epic mini-games
-
Jet-setting travel options
-
-
You can also interact with Angela by talking to her, stroking her, poking her, or making her smile. She will respond with her own voice and expressions. You can even record videos of your interactions and share them with your friends.
-
How to download My Talking Angela APK iOS?
-
The official way: App Store
-
The easiest and safest way to download My Talking Angela APK iOS is through the App Store. You can simply search for the game on the App Store or click on this link to go directly to the download page. The game is free to download and play, but it offers in-app purchases for extra coins, diamonds, and items. You can also subscribe to get exclusive benefits such as unlimited energy, double coins, no ads, and more.
-
The alternative way: Internet Archive
-
If you want to try out the older versions of My Talking Angela APK iOS, you can use the Internet Archive website. This website archives various digital content, including apps, games, books, music, videos, and more. You can find the first version of My Talking Angela APK iOS from 2014 on this link. However, this version is only compatible with iOS 6.0 to iOS 6.1.6 devices such as iPad, iPhone, and iPod touch. You will need to use an emulator or a jailbroken device to run this version.
-
How to play My Talking Angela APK iOS?
-
Customize Angela's appearance and outfits
-
One of the main attractions of My Talking Angela APK iOS is that you can customize Angela's appearance and outfits according to your preferences. You can change her fur color, eye color, makeup style, hairstyle, accessories, clothes, shoes, and more. You can also unlock new items by leveling up or buying them with coins or diamonds.
-
Interact with Angela and make her happy
-
Another important aspect of My Talking Angela APK iOS is that you can interact with Angela and make her happy. You can talk to her and she will repeat what you say in a cute voice. You can also touch her, tickle her, or make funny faces at her. She will react with different emotions and sounds. You can also feed her, bathe her, brush her teeth, and put her to bed. You need to take care of her basic needs such as hunger, hygiene, energy, and happiness. If you neglect her, she will become sad or sick.
-
My Talking Angela 2 app download for ios
-How to install My Talking Angela apk on iphone
-My Talking Angela virtual pet game for ipad
-My Talking Angela 1.0 long lost version for ios 6
-My Talking Angela fashion star dress up game for ios
-My Talking Angela apk mod unlimited money for ios
-My Talking Angela free stickers and rewards for ios
-My Talking Angela baking and dancing activities for ios
-My Talking Angela spring magic update for ios
-My Talking Angela rumors and myths debunked for ios
-My Talking Angela best friend and chatbot for ios
-My Talking Angela youtube videos and songs for ios
-My Talking Angela customer support and feedback for ios
-My Talking Angela privacy policy and terms of use for ios
-My Talking Angela subscription and in-app purchases for ios
-My Talking Angela mini-games and puzzles for ios
-My Talking Angela jet-setting travel options for ios
-My Talking Angela makeup and hair salon for ios
-My Talking Angela wardrobe and fashion choices for ios
-My Talking Angela super fun virtual star for ios
-My Talking Angela 3D world and graphics for ios
-My Talking Angela offline mode and data usage for ios
-My Talking Angela tips and tricks for beginners for ios
-My Talking Angela cheats and hacks for advanced players for ios
-My Talking Angela reviews and ratings from users for ios
-My Talking Angela compatible devices and versions for ios
-My Talking Angela size and storage space required for ios
-My Talking Angela latest news and updates for ios
-My Talking Angela social media and community for ios
-My Talking Angela screenshots and wallpapers for ios
-How to uninstall or delete My Talking Angela apk from ios
-How to backup or restore My Talking Angela data on ios
-How to sync or transfer My Talking Angela progress across devices on ios
-How to fix or troubleshoot My Talking Angela issues on ios
-How to contact or report My Talking Angela developers on ios
-How to customize or personalize My Talking Angela settings on ios
-How to enable or disable notifications from My Talking Angela on ios
-How to earn or spend coins and diamonds in My Talking Angela on ios
-How to level up or unlock new features in My Talking Angela on ios
-How to play or interact with My Talking Angela on ios
-
Explore different locations and mini-games
-
My Talking Angela APK iOS also offers various locations and mini-games for you to explore and enjoy. You can visit Angela's cozy home, where you can play with her toys, watch TV, or listen to music. You can also travel to different places such as Paris, New York, Tokyo, or the beach. Each location has its own theme and activities. You can also play fun mini-games with Angela, such as Happy Connect, Bubble Shooter, Brick Breaker, and more. These games will help you earn coins and diamonds, as well as improve Angela's skills and mood.
-
Tips and tricks for My Talking Angela APK iOS
-
Collect coins and diamonds
-
Coins and diamonds are the main currencies in My Talking Angela APK iOS. You can use them to buy new items, unlock new locations, or access premium features. You can earn coins and diamonds by playing mini-games, completing daily tasks and achievements, watching videos and ads, or using real money. You can also get free coins and diamonds by logging in every day, spinning the wheel of fortune, or opening mystery boxes.
-
Complete daily tasks and achievements
-
Daily tasks and achievements are another way to earn coins and diamonds in My Talking Angela APK iOS. Daily tasks are simple actions that you need to do every day, such as feeding Angela, playing with her, or dressing her up. Achievements are more challenging goals that you need to accomplish over time, such as collecting a certain number of stickers, reaching a certain level, or playing a certain mini-game. Completing these tasks and achievements will not only reward you with coins and diamonds, but also with experience points that will help you level up faster.
-
Watch videos and ads for rewards
-
Watching videos and ads is another option to get free rewards in My Talking Angela APK iOS. You can watch videos and ads by tapping on the TV icon in Angela's home or by clicking on the offer wall in the shop. You can get various rewards such as coins, diamonds, energy refills, stickers, or mystery boxes. However, you need to have an internet connection to watch videos and ads.
-
Conclusion
-
My Talking Angela APK iOS is a fun and cute virtual pet game that will keep you entertained for hours. You can adopt Angela as your own pet and take care of her as she grows from a kitten to a cat. You can customize her appearance and outfits, interact with her and make her happy, explore different locations and mini-games, and collect coins and diamonds. You can also record videos of your interactions and share them with your friends. If you are looking for a game that combines simulation, fashion, adventure, and fun, you should download My Talking Angela APK iOS today.
-
FAQs
-
-
Q: Is My Talking Angela APK iOS safe for kids?
-
A: My Talking Angela APK iOS is rated 4+ on the App Store, which means it is suitable for everyone. However, some parents may have concerns about the game's privacy policy, in-app purchases, and online interactions. Therefore, we recommend that parents supervise their kids while playing the game and use the parental control settings to limit or disable certain features.
-
Q: How can I get more stickers in My Talking Angela APK iOS?
-
A: Stickers are collectible items that you can find in mystery boxes or by traveling to different locations. You can also trade stickers with other players by using the sticker album feature. You need to have an internet connection and a Facebook account to use this feature. You can also buy stickers with diamonds in the shop.
-
Q: How can I backup or restore my progress in My Talking Angela APK iOS?
-
A: You can backup or restore your progress in My Talking Angela APK iOS by using the cloud save feature. You need to have an internet connection and a Facebook account to use this feature. You can access the cloud save feature by tapping on the settings icon in the top right corner of the screen and then tapping on the cloud icon. You can then choose to upload or download your progress.
-
Q: How can I change the language in My Talking Angela APK iOS?
-
A: You can change the language in My Talking Angela APK iOS by tapping on the settings icon in the top right corner of the screen and then tapping on the language icon. You can then choose from 32 different languages, including English, Spanish, French, German, Chinese, Japanese, and more.
-
Q: How can I contact the developers of My Talking Angela APK iOS?
-
A: You can contact the developers of My Talking Angela APK iOS by tapping on the settings icon in the top right corner of the screen and then tapping on the support icon. You can then choose to send an email, visit the website, or follow them on social media.
-
197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Piano Magic Tiles Play Music and Enjoy Mod APK Download.md b/spaces/congsaPfin/Manga-OCR/logs/Piano Magic Tiles Play Music and Enjoy Mod APK Download.md
deleted file mode 100644
index 39f5d964586339d3e8a44b6097e7ed7bb290ab6d..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Piano Magic Tiles Play Music and Enjoy Mod APK Download.md
+++ /dev/null
@@ -1,77 +0,0 @@
-
-
Piano Music Tiles: Magic Tiles Mod APK - A Fun and Relaxing Game for Music Lovers
-
Do you love music and piano? Do you want to play your favorite songs on your mobile device? Do you want to have a fun and relaxing time with a simple and addictive game? If you answered yes to any of these questions, then you should try Piano Music Tiles: Magic Tiles, a music piano game that will test your rhythm and reflexes. And if you want to have an even better experience, you should download the mod apk version of this game, which will give you unlimited coins and diamonds, no ads and pop-ups, and access to all songs and themes. In this article, we will tell you everything you need to know about Piano Music Tiles: Magic Tiles mod apk, including what it is, how to play it, why you should download it, what features it has, and how to download and install it.
Piano Music Tiles: Magic Tiles is a music piano game that was developed by Piano Magic Tiles Challenge Music Free. It is available for Android devices on Google Play Store. The game has over 10 million downloads and a 4.5-star rating. The game is simple but challenging: all you need to do is feel the music and tap the black tiles. But remember, don't touch the white tiles or you will lose. The game has various types of music and genres, such as classical, pop, rock, EDM, anime, etc. You can also choose from different themes and backgrounds to customize your game. The game also has a challenge mode where you can compete with other players around the world on the leaderboard. You can also play offline and save your progress on the cloud.
-
Why download Piano Music Tiles: Magic Tiles mod apk?
-
While Piano Music Tiles: Magic Tiles is a free game, it also has some limitations and drawbacks that can affect your enjoyment. For example, you need coins and diamonds to unlock new songs and themes, but they are hard to earn and expensive to buy. You also have to deal with annoying ads and pop-ups that interrupt your game. And some songs and themes are locked behind a paywall or require a subscription. That's why we recommend downloading Piano Music Tiles: Magic Tiles mod apk, which will give you the following benefits:
-
Unlimited coins and diamonds
-
With Piano Music Tiles: Magic Tiles mod apk, you will have unlimited coins and diamonds in your account. You can use them to unlock any song or theme you want without spending real money. You can also use them to buy boosters and power-ups that will help you improve your score and performance.
-
No ads and pop-ups
-
With Piano Music Tiles: Magic Tiles mod apk, you will not see any ads or pop-ups on your screen. You can enjoy the game without any interruption or distraction. You can also save your data and battery life by not loading any unnecessary content.
-
Unlock all songs and themes
-
With Piano Music Tiles: Magic Tiles mod apk, you will have access to all songs and themes in the game. You can play any song or genre you like without waiting for it to be unlocked or paying for it. You can also choose from different themes and backgrounds to suit your mood and preference.
-
Features of Piano Music Tiles
Features of Piano Music Tiles: Magic Tiles
-
Piano Music Tiles: Magic Tiles is not just a simple music piano game. It also has many features that make it more fun and relaxing. Here are some of the features that you can enjoy with this game:
-
piano magic tiles music game mod apk
-magic tiles piano music challenge mod apk
-piano music tiles free mod apk
-magic tiles piano music songs mod apk
-piano magic tiles pop music mod apk
-magic tiles piano music master mod apk
-piano music tiles 2 mod apk
-magic tiles piano music anime mod apk
-piano magic tiles classic music mod apk
-magic tiles piano music rock mod apk
-piano music tiles 3 mod apk
-magic tiles piano music edm mod apk
-piano magic tiles kpop music mod apk
-magic tiles piano music offline mod apk
-piano music tiles 4 mod apk
-magic tiles piano music online mod apk
-piano magic tiles christmas music mod apk
-magic tiles piano music kids mod apk
-piano music tiles 5 mod apk
-magic tiles piano music relax mod apk
-piano magic tiles bts music mod apk
-magic tiles piano music quiz mod apk
-piano music tiles 6 mod apk
-magic tiles piano music maker mod apk
-piano magic tiles jazz music mod apk
-magic tiles piano music simulator mod apk
-piano music tiles 7 mod apk
-magic tiles piano music tutorial mod apk
-piano magic tiles rap music mod apk
-magic tiles piano music converter mod apk
-piano music tiles 8 mod apk
-magic tiles piano music downloader mod apk
-piano magic tiles country music mod apk
-magic tiles piano music editor mod apk
-piano music tiles 9 mod apk
-magic tiles piano music generator mod apk
-piano magic tiles disco music mod apk
-magic tiles piano music recorder mod apk
-piano music tiles 10 mod apk
-magic tiles piano music player mod apk
-
Various types of music and genres
-
Piano Music Tiles: Magic Tiles has a huge collection of songs and music that you can play. You can find songs from different genres, such as classical, pop, rock, EDM, anime, etc. You can also find songs from famous artists, such as Beethoven, Mozart, Chopin, Taylor Swift, Ed Sheeran, BTS, etc. You can also request new songs and genres from the developers. You will never get bored with this game because there is always something new to play.
-
Relaxing visual design and sound effects
-
Piano Music Tiles: Magic Tiles has a relaxing visual design and sound effects that will make you feel calm and peaceful. The game has different themes and backgrounds that you can choose from, such as night sky, forest, ocean, etc. The game also has realistic piano sound effects that will make you feel like you are playing a real piano. The game is designed to help you relax and enjoy the music.
-
Challenge mode and leaderboard
-
Piano Music Tiles: Magic Tiles also has a challenge mode where you can test your skills and compete with other players around the world. The challenge mode has different levels of difficulty and speed that you can choose from. The game also has a leaderboard where you can see your rank and score among other players. You can also share your achievements and progress with your friends on social media. The game is designed to challenge you and motivate you to improve.
-
Offline mode and cloud save
-
Piano Music Tiles: Magic Tiles also has an offline mode where you can play the game without an internet connection. You can play any song or theme that you have unlocked or downloaded without any limitation. The game also has a cloud save feature where you can save your progress and data on the cloud. You can access your account and data on any device that you use. The game is designed to be convenient and accessible.
-
How to download and install Piano Music Tiles: Magic Tiles mod apk?
-
If you are interested in downloading and installing Piano Music Tiles: Magic Tiles mod apk, you can follow these simple steps:
-
Step 1: Download the mod apk file from a trusted source
-
The first step is to download the mod apk file from a trusted source. You can find many websites that offer mod apk files for various games and apps, but not all of them are safe and reliable. Some of them may contain viruses or malware that can harm your device or steal your data. That's why we recommend downloading the mod apk file from our website, which is 100% safe and secure. You can download the mod apk file by clicking on this link: [Piano Music Tiles: Magic Tiles Mod APK].
-
Step 2: Enable unknown sources on your device settings
-
The second step is to enable unknown sources on your device settings. This is necessary because Android devices do not allow installing apps from sources other than Google Play Store by default. To enable unknown sources, you need to go to your device settings > security > unknown sources > toggle on. This will allow you to install apps from sources other than Google Play Store.
-
Step 3: Install the mod apk file and enjoy the game
-
The third step is to install the mod apk file and enjoy the game. To install the mod apk file, you need to locate the file on your device storage > tap on it > follow the instructions on the screen > wait for the installation to finish > open the game and enjoy. You will see that you have unlimited coins and diamonds, no ads and pop-ups, and access to all songs and themes in the game.
-
Conclusion
-
Piano Music Tiles: Magic Tiles is a music piano game that will give you a fun and relaxing time with your favorite songs. You can play various types of music and genres, customize your game with different themes and backgrounds, compete with other players on the challenge mode and leaderboard, and play offline and save your progress on the cloud. And if you want to have an even better experience, you should download Piano Music Tiles: Magic Tiles mod apk, which will give you unlimited coins and diamonds, no ads and pop-ups, and access to all songs and themes in the game. Download Piano Music Tiles: Magic Tiles mod apk now and enjoy the music!
- FAQs Q: Is Piano Music Tiles: Magic Tiles mod apk safe to use? A: A: Yes, Piano Music Tiles: Magic Tiles mod apk is safe to use. It does not contain any viruses or malware that can harm your device or steal your data. It also does not require any root or jailbreak to install or run. However, you should always download the mod apk file from a trusted source, such as our website, to avoid any risks. Q: How can I update Piano Music Tiles: Magic Tiles mod apk? A: To update Piano Music Tiles: Magic Tiles mod apk, you need to download the latest version of the mod apk file from our website and install it over the existing one. You do not need to uninstall the previous version or lose your progress. However, you should always backup your data before updating to avoid any issues. Q: Can I play Piano Music Tiles: Magic Tiles mod apk with my friends? A: Yes, you can play Piano Music Tiles: Magic Tiles mod apk with your friends. You can connect your game with your Facebook account and invite your friends to join you. You can also see your friends' scores and achievements on the leaderboard and challenge them to beat you. Q: What are the minimum requirements to play Piano Music Tiles: Magic Tiles mod apk? A: The minimum requirements to play Piano Music Tiles: Magic Tiles mod apk are: - Android 4.1 or higher - 50 MB of free storage space - Internet connection (optional) Q: How can I contact the developers of Piano Music Tiles: Magic Tiles mod apk? A: If you have any questions, feedback, or suggestions for Piano Music Tiles: Magic Tiles mod apk, you can contact the developers by emailing them at [pianomagictileschallenge@gmail.com] or visiting their Facebook page at [Piano Magic Tiles Challenge Music Free]. 401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Unlimited Money Weapons and Vehicles in Rope Hero Mafia City Wars Hack APK.md b/spaces/congsaPfin/Manga-OCR/logs/Unlimited Money Weapons and Vehicles in Rope Hero Mafia City Wars Hack APK.md
deleted file mode 100644
index a08f94de4a338ae6f1ee807d3a6edd5f2d722a50..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Unlimited Money Weapons and Vehicles in Rope Hero Mafia City Wars Hack APK.md
+++ /dev/null
@@ -1,128 +0,0 @@
-
-
Rope Hero: Mafia City Wars Hack APK - How to Get Unlimited Money and Diamonds
-
Are you a fan of superhero games? Do you want to become a super rope hero who can fight crime and save the city? If yes, then you should try Rope Hero: Mafia City Wars, a thrilling action game with RPG elements. In this game, you can use your superpowers and guns to fight with the gangsters, capture districts, and complete quests. You can also customize your super rope hero with different skins and weapons.
-
However, to enjoy the game fully, you will need a lot of money and diamonds. Money is used to buy weapons, vehicles, and upgrades, while diamonds are used to unlock premium skins and items. Earning money and diamonds in the game is not easy, as you have to complete missions, watch ads, or spend real money. That's why many players are looking for a way to get unlimited money and diamonds in Rope Hero: Mafia City Wars.
Fortunately, there is a solution for that. You can use a hack apk, which is a modified version of the original game that gives you access to unlimited resources. With a hack apk, you can enjoy the game without any limitations or restrictions. You can buy anything you want, unlock everything you need, and have more fun playing Rope Hero: Mafia City Wars.
-
Features of Rope Hero: Mafia City Wars Hack APK
-
A hack apk is not just a simple cheat tool. It is a fully functional game that has been modified to provide you with some amazing features that are not available in the original game. Here are some of the features of Rope Hero: Mafia City Wars Hack APK:
-
Unlimited money and diamonds
-
This is the main feature of the hack apk. You will get unlimited money and diamonds in your account as soon as you install the hack apk. You can use them to buy anything you want in the game, such as weapons, vehicles, upgrades, skins, and items. You don't have to worry about running out of money or diamonds ever again.
-
Unlock all superhero skins and weapons
-
Another feature of the hack apk is that it unlocks all the superhero skins and weapons in the game. You can choose from a variety of skins for your super rope hero, such as Spider-Man, Iron Man, Batman, Hulk, Deadpool, and more. You can also equip your hero with different weapons, such as pistols, rifles, shotguns, rocket launchers, grenades, swords, axes, hammers, and more. You can mix and match different skins and weapons to create your own unique superhero.
-
No ads and no root required
-
The hack apk also removes all the annoying ads and pop-ups that interrupt your gameplay. You can enjoy the game without any distractions or interruptions. The hack apk also does not require root access to work. You can install it on any Android device without worrying about rooting your device or voiding your warranty.
-
How to Download and Install Rope Hero: Mafia City Wars Hack APK
-
Downloading and installing the hack apk is very easy and simple. You just need to follow these steps:
-
Step 1: Enable unknown sources on your device
-
Before you can install the hack apk, you need to enable unknown sources on your device. This will allow you to install apps from sources other than the Google Play Store. To do this, go to your device settings, then security, then unknown sources, and turn it on.
-
Step 2: Download the hack apk file from a trusted source
-
Next, you need to download the hack apk file from a trusted source. You can use the link below to download the latest version of Rope Hero: Mafia City Wars Hack APK. The file size is about 100 MB, so make sure you have enough space on your device.
-
rope hero mafia city wars mod apk unlimited money
-rope hero mafia city wars cheats android
-rope hero mafia city wars hack download
-rope hero mafia city wars game online
-rope hero mafia city wars apk free
-rope hero mafia city wars mod menu
-rope hero mafia city wars unlimited gems
-rope hero mafia city wars latest version
-rope hero mafia city wars gameplay
-rope hero mafia city wars hack ios
-rope hero mafia city wars no ads
-rope hero mafia city wars tips and tricks
-rope hero mafia city wars review
-rope hero mafia city wars best weapons
-rope hero mafia city wars offline
-rope hero mafia city wars hack tool
-rope hero mafia city wars for pc
-rope hero mafia city wars all characters
-rope hero mafia city wars guide
-rope hero mafia city wars codes
-rope hero mafia city wars mod apk revdl
-rope hero mafia city wars hack apk 2023
-rope hero mafia city wars new update
-rope hero mafia city wars superpowers
-rope hero mafia city wars how to play
-rope hero mafia city wars hack apk an1.com[^1^]
-rope hero mafia city wars missions
-rope hero mafia city wars secrets
-rope hero mafia city wars vehicles
-rope hero mafia city wars hack apk happymod
-rope hero mafia city wars android 1
-rope hero mafia city wars mod apk rexdl
-rope hero mafia city wars hack version
-rope hero mafia city wars download for android
-rope hero mafia city wars mod apk android 1
-rope hero mafia city wars hack no verification
-rope hero mafia city wars wiki
-rope hero mafia city wars mod apk 2023
-rope hero mafia city wars hack online generator
-rope hero mafia city wars unlimited everything
-rope hero mafia city wars mod apk latest version download
-rope hero mafia city wars hack without human verification
-rope hero mafia city wars free gems and coins
-rope hero mafia city wars mod apk obb download
Step 3: Install the hack apk file and launch the game
-
Finally, you need to install the hack apk file and launch the game. To do this, locate the downloaded file on your device, tap on it, and follow the instructions on the screen. Once the installation is complete, open the game and enjoy unlimited money and diamonds.
-
Tips and Tricks for Playing Rope Hero: Mafia City Wars
-
Rope Hero: Mafia City Wars is a fun and addictive game that will keep you entertained for hours. However, if you want to master the game and become the best super rope hero in the city, you will need some tips and tricks. Here are some of them:
-
Use your superpowers wisely
-
Your superpowers are your main weapons in the game. You can use them to swing around the city, climb buildings, jump over obstacles, and fight enemies. However, you should also be careful not to overuse them, as they consume energy. You can replenish your energy by collecting blue orbs or using money or diamonds.
-
Explore the open world and complete quests
-
The game features a large open world that you can explore freely. You can find various locations, such as shops, banks, casinos, police stations, hospitals, and more. You can also interact with different characters, such as civilians, gangsters, cops, and superheroes. You can also complete various quests that will reward you with money, diamonds, experience points, and items. Quests are marked with yellow icons on the map.
-
Fight with the gangster bosses and capture districts
-
The city is divided into several districts that are controlled by different gangster bosses. You can challenge them to a fight and try to capture their districts. This will increase your reputation and influence in the city. You can also earn more money and diamonds by collecting taxes from the captured districts. However, be prepared to face strong resistance from the gangsters and their minions.
-
Conclusion
-
Rope Hero: Mafia City Wars is an exciting game that lets you become a super rope hero who can save the city from crime and chaos. You can use your superpowers and weapons to fight with the gangsters, capture districts, and complete quests. You can also customize your super rope hero with different skins and weapons.
-
If you want to enjoy the game without any limitations or restrictions, you can use a hack apk that gives you unlimited money and diamonds. With a hack apk, you can unlock everything you need in the game and have more fun playing Rope Hero: Mafia City Wars.
-
So what are you waiting for? Download Rope Hero: Mafia City Wars Hack APK now and become the ultimate super rope hero in the city!
-
FAQs
-
Is Rope Hero: Mafia City Wars Hack APK safe to use?
-
Yes, Rope Hero: Mafia City Wars Hack APK is safe to use. It does not contain any viruses or malware that can harm your device or compromise your privacy. However, you should always download it from a trusted source and scan it with an antivirus before installing it.
-
Will I get banned for using Rope Hero: Mafia City Wars Hack APK?
-
No, you will not get banned for using Rope Hero: Mafia City Wars Hack APK. The hack apk is undetectable by the game servers and does not interfere with other players' gameplay. However, you should avoid using it excessively or in a way that affects other players' enjoyment of the game. You should also respect the game rules and terms of service.
-
How can I update Rope Hero: Mafia City Wars Hack APK?
-
To update Rope Hero: Mafia City Wars Hack APK, you need to download the latest version of the hack apk from the same source you downloaded it from before. You can check the version number and the date of the hack apk on the download page. You can also follow the updates and news of the hack apk on its official website or social media pages. To install the update, you need to uninstall the previous version of the hack apk and install the new one.
-
What are the best superhero skins and weapons in Rope Hero: Mafia City Wars?
-
The best superhero skins and weapons in Rope Hero: Mafia City Wars depend on your personal preference and play style. However, some of the most popular and powerful ones are:
-
-
-
Skin
-
Weapon
-
Description
-
-
-
Spider-Man
-
Web Shooter
-
A classic superhero skin that lets you swing around the city with your web shooter. You can also shoot webs at enemies to immobilize them or pull them towards you.
-
-
-
Iron Man
-
Repulsor Blast
-
A futuristic superhero skin that gives you a suit of armor with jet boosters and repulsor blasts. You can fly around the city and blast enemies with your powerful beams.
-
-
-
Batman
-
Batarang
-
A dark and mysterious superhero skin that gives you a cape and a batarang. You can glide around the city and throw batarangs at enemies to stun them or knock them out.
-
-
-
Hulk
-
Fists
-
A monstrous superhero skin that gives you incredible strength and durability. You can smash enemies with your fists or throw objects at them. You can also jump high and cause shockwaves when you land.
-
-
-
Deadpool
-
Dual Swords
-
A humorous and sarcastic superhero skin that gives you dual swords and a healing factor. You can slash enemies with your swords or use them to deflect bullets. You can also heal from any damage quickly.
-
-
-
How can I contact the developer of Rope Hero: Mafia City Wars?
-
If you have any questions, feedback, suggestions, or issues regarding Rope Hero: Mafia City Wars, you can contact the developer of the game through their email address or their social media pages. Here are their contact details:
-
-January 12, 2015 - staralbu 7b17bfd26b staralbu February 15 at 00:32. And what to do when people in the villages with children cannot go to the cities?
-But what about young children if they do not have the opportunity to go to the clinic?
-But what if people in the village have a car breakdown and there is no gasoline in the hospital?
-And how to live for people who cannot buy their own medicines?
-These are not just words.
-Before my eyes, when I was working in an orphanage, a girl died.
-She didn't have parents.
-We treated her.
-She had a very advanced form of hepatitis B. The girl was doomed to die. 8a78ff9644
-
-
-
diff --git a/spaces/contluForse/HuggingGPT/assets/CopyTrans Control Center Crack ((TOP)).rar.md b/spaces/contluForse/HuggingGPT/assets/CopyTrans Control Center Crack ((TOP)).rar.md
deleted file mode 100644
index 5331b6224c1d4011905e5dec4306a3fb8f58b16c..0000000000000000000000000000000000000000
--- a/spaces/contluForse/HuggingGPT/assets/CopyTrans Control Center Crack ((TOP)).rar.md
+++ /dev/null
@@ -1,6 +0,0 @@
-
This demo is strictly for research demo purpose only. For commercial use please contact us.
-
- """
- )
-
-iface.queue(concurrency_count = 2)
-iface.launch(debug=True)
-# iface.launch(debug=True, share=True)
\ No newline at end of file
diff --git a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/_textwrap.py b/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/_textwrap.py
deleted file mode 100644
index b47dcbd4264e86715adfae1c5124c288b67a983e..0000000000000000000000000000000000000000
--- a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/_textwrap.py
+++ /dev/null
@@ -1,49 +0,0 @@
-import textwrap
-import typing as t
-from contextlib import contextmanager
-
-
-class TextWrapper(textwrap.TextWrapper):
- def _handle_long_word(
- self,
- reversed_chunks: t.List[str],
- cur_line: t.List[str],
- cur_len: int,
- width: int,
- ) -> None:
- space_left = max(width - cur_len, 1)
-
- if self.break_long_words:
- last = reversed_chunks[-1]
- cut = last[:space_left]
- res = last[space_left:]
- cur_line.append(cut)
- reversed_chunks[-1] = res
- elif not cur_line:
- cur_line.append(reversed_chunks.pop())
-
- @contextmanager
- def extra_indent(self, indent: str) -> t.Iterator[None]:
- old_initial_indent = self.initial_indent
- old_subsequent_indent = self.subsequent_indent
- self.initial_indent += indent
- self.subsequent_indent += indent
-
- try:
- yield
- finally:
- self.initial_indent = old_initial_indent
- self.subsequent_indent = old_subsequent_indent
-
- def indent_only(self, text: str) -> str:
- rv = []
-
- for idx, line in enumerate(text.splitlines()):
- indent = self.initial_indent
-
- if idx > 0:
- indent = self.subsequent_indent
-
- rv.append(f"{indent}{line}")
-
- return "\n".join(rv)
diff --git a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/types.py b/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/types.py
deleted file mode 100644
index 2b1d1797f2e115e9bc976bcaf7d8e1884a91e91c..0000000000000000000000000000000000000000
--- a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/click/types.py
+++ /dev/null
@@ -1,1089 +0,0 @@
-import os
-import stat
-import sys
-import typing as t
-from datetime import datetime
-from gettext import gettext as _
-from gettext import ngettext
-
-from ._compat import _get_argv_encoding
-from ._compat import open_stream
-from .exceptions import BadParameter
-from .utils import format_filename
-from .utils import LazyFile
-from .utils import safecall
-
-if t.TYPE_CHECKING:
- import typing_extensions as te
- from .core import Context
- from .core import Parameter
- from .shell_completion import CompletionItem
-
-
-class ParamType:
- """Represents the type of a parameter. Validates and converts values
- from the command line or Python into the correct type.
-
- To implement a custom type, subclass and implement at least the
- following:
-
- - The :attr:`name` class attribute must be set.
- - Calling an instance of the type with ``None`` must return
- ``None``. This is already implemented by default.
- - :meth:`convert` must convert string values to the correct type.
- - :meth:`convert` must accept values that are already the correct
- type.
- - It must be able to convert a value if the ``ctx`` and ``param``
- arguments are ``None``. This can occur when converting prompt
- input.
- """
-
- is_composite: t.ClassVar[bool] = False
- arity: t.ClassVar[int] = 1
-
- #: the descriptive name of this type
- name: str
-
- #: if a list of this type is expected and the value is pulled from a
- #: string environment variable, this is what splits it up. `None`
- #: means any whitespace. For all parameters the general rule is that
- #: whitespace splits them up. The exception are paths and files which
- #: are split by ``os.path.pathsep`` by default (":" on Unix and ";" on
- #: Windows).
- envvar_list_splitter: t.ClassVar[t.Optional[str]] = None
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- """Gather information that could be useful for a tool generating
- user-facing documentation.
-
- Use :meth:`click.Context.to_info_dict` to traverse the entire
- CLI structure.
-
- .. versionadded:: 8.0
- """
- # The class name without the "ParamType" suffix.
- param_type = type(self).__name__.partition("ParamType")[0]
- param_type = param_type.partition("ParameterType")[0]
-
- # Custom subclasses might not remember to set a name.
- if hasattr(self, "name"):
- name = self.name
- else:
- name = param_type
-
- return {"param_type": param_type, "name": name}
-
- def __call__(
- self,
- value: t.Any,
- param: t.Optional["Parameter"] = None,
- ctx: t.Optional["Context"] = None,
- ) -> t.Any:
- if value is not None:
- return self.convert(value, param, ctx)
-
- def get_metavar(self, param: "Parameter") -> t.Optional[str]:
- """Returns the metavar default for this param if it provides one."""
-
- def get_missing_message(self, param: "Parameter") -> t.Optional[str]:
- """Optionally might return extra information about a missing
- parameter.
-
- .. versionadded:: 2.0
- """
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- """Convert the value to the correct type. This is not called if
- the value is ``None`` (the missing value).
-
- This must accept string values from the command line, as well as
- values that are already the correct type. It may also convert
- other compatible types.
-
- The ``param`` and ``ctx`` arguments may be ``None`` in certain
- situations, such as when converting prompt input.
-
- If the value cannot be converted, call :meth:`fail` with a
- descriptive message.
-
- :param value: The value to convert.
- :param param: The parameter that is using this type to convert
- its value. May be ``None``.
- :param ctx: The current context that arrived at this value. May
- be ``None``.
- """
- return value
-
- def split_envvar_value(self, rv: str) -> t.Sequence[str]:
- """Given a value from an environment variable this splits it up
- into small chunks depending on the defined envvar list splitter.
-
- If the splitter is set to `None`, which means that whitespace splits,
- then leading and trailing whitespace is ignored. Otherwise, leading
- and trailing splitters usually lead to empty items being included.
- """
- return (rv or "").split(self.envvar_list_splitter)
-
- def fail(
- self,
- message: str,
- param: t.Optional["Parameter"] = None,
- ctx: t.Optional["Context"] = None,
- ) -> "t.NoReturn":
- """Helper method to fail with an invalid value message."""
- raise BadParameter(message, ctx=ctx, param=param)
-
- def shell_complete(
- self, ctx: "Context", param: "Parameter", incomplete: str
- ) -> t.List["CompletionItem"]:
- """Return a list of
- :class:`~click.shell_completion.CompletionItem` objects for the
- incomplete value. Most types do not provide completions, but
- some do, and this allows custom types to provide custom
- completions as well.
-
- :param ctx: Invocation context for this command.
- :param param: The parameter that is requesting completion.
- :param incomplete: Value being completed. May be empty.
-
- .. versionadded:: 8.0
- """
- return []
-
-
-class CompositeParamType(ParamType):
- is_composite = True
-
- @property
- def arity(self) -> int: # type: ignore
- raise NotImplementedError()
-
-
-class FuncParamType(ParamType):
- def __init__(self, func: t.Callable[[t.Any], t.Any]) -> None:
- self.name: str = func.__name__
- self.func = func
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict["func"] = self.func
- return info_dict
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- try:
- return self.func(value)
- except ValueError:
- try:
- value = str(value)
- except UnicodeError:
- value = value.decode("utf-8", "replace")
-
- self.fail(value, param, ctx)
-
-
-class UnprocessedParamType(ParamType):
- name = "text"
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- return value
-
- def __repr__(self) -> str:
- return "UNPROCESSED"
-
-
-class StringParamType(ParamType):
- name = "text"
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- if isinstance(value, bytes):
- enc = _get_argv_encoding()
- try:
- value = value.decode(enc)
- except UnicodeError:
- fs_enc = sys.getfilesystemencoding()
- if fs_enc != enc:
- try:
- value = value.decode(fs_enc)
- except UnicodeError:
- value = value.decode("utf-8", "replace")
- else:
- value = value.decode("utf-8", "replace")
- return value
- return str(value)
-
- def __repr__(self) -> str:
- return "STRING"
-
-
-class Choice(ParamType):
- """The choice type allows a value to be checked against a fixed set
- of supported values. All of these values have to be strings.
-
- You should only pass a list or tuple of choices. Other iterables
- (like generators) may lead to surprising results.
-
- The resulting value will always be one of the originally passed choices
- regardless of ``case_sensitive`` or any ``ctx.token_normalize_func``
- being specified.
-
- See :ref:`choice-opts` for an example.
-
- :param case_sensitive: Set to false to make choices case
- insensitive. Defaults to true.
- """
-
- name = "choice"
-
- def __init__(self, choices: t.Sequence[str], case_sensitive: bool = True) -> None:
- self.choices = choices
- self.case_sensitive = case_sensitive
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict["choices"] = self.choices
- info_dict["case_sensitive"] = self.case_sensitive
- return info_dict
-
- def get_metavar(self, param: "Parameter") -> str:
- choices_str = "|".join(self.choices)
-
- # Use curly braces to indicate a required argument.
- if param.required and param.param_type_name == "argument":
- return f"{{{choices_str}}}"
-
- # Use square braces to indicate an option or optional argument.
- return f"[{choices_str}]"
-
- def get_missing_message(self, param: "Parameter") -> str:
- return _("Choose from:\n\t{choices}").format(choices=",\n\t".join(self.choices))
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- # Match through normalization and case sensitivity
- # first do token_normalize_func, then lowercase
- # preserve original `value` to produce an accurate message in
- # `self.fail`
- normed_value = value
- normed_choices = {choice: choice for choice in self.choices}
-
- if ctx is not None and ctx.token_normalize_func is not None:
- normed_value = ctx.token_normalize_func(value)
- normed_choices = {
- ctx.token_normalize_func(normed_choice): original
- for normed_choice, original in normed_choices.items()
- }
-
- if not self.case_sensitive:
- normed_value = normed_value.casefold()
- normed_choices = {
- normed_choice.casefold(): original
- for normed_choice, original in normed_choices.items()
- }
-
- if normed_value in normed_choices:
- return normed_choices[normed_value]
-
- choices_str = ", ".join(map(repr, self.choices))
- self.fail(
- ngettext(
- "{value!r} is not {choice}.",
- "{value!r} is not one of {choices}.",
- len(self.choices),
- ).format(value=value, choice=choices_str, choices=choices_str),
- param,
- ctx,
- )
-
- def __repr__(self) -> str:
- return f"Choice({list(self.choices)})"
-
- def shell_complete(
- self, ctx: "Context", param: "Parameter", incomplete: str
- ) -> t.List["CompletionItem"]:
- """Complete choices that start with the incomplete value.
-
- :param ctx: Invocation context for this command.
- :param param: The parameter that is requesting completion.
- :param incomplete: Value being completed. May be empty.
-
- .. versionadded:: 8.0
- """
- from click.shell_completion import CompletionItem
-
- str_choices = map(str, self.choices)
-
- if self.case_sensitive:
- matched = (c for c in str_choices if c.startswith(incomplete))
- else:
- incomplete = incomplete.lower()
- matched = (c for c in str_choices if c.lower().startswith(incomplete))
-
- return [CompletionItem(c) for c in matched]
-
-
-class DateTime(ParamType):
- """The DateTime type converts date strings into `datetime` objects.
-
- The format strings which are checked are configurable, but default to some
- common (non-timezone aware) ISO 8601 formats.
-
- When specifying *DateTime* formats, you should only pass a list or a tuple.
- Other iterables, like generators, may lead to surprising results.
-
- The format strings are processed using ``datetime.strptime``, and this
- consequently defines the format strings which are allowed.
-
- Parsing is tried using each format, in order, and the first format which
- parses successfully is used.
-
- :param formats: A list or tuple of date format strings, in the order in
- which they should be tried. Defaults to
- ``'%Y-%m-%d'``, ``'%Y-%m-%dT%H:%M:%S'``,
- ``'%Y-%m-%d %H:%M:%S'``.
- """
-
- name = "datetime"
-
- def __init__(self, formats: t.Optional[t.Sequence[str]] = None):
- self.formats: t.Sequence[str] = formats or [
- "%Y-%m-%d",
- "%Y-%m-%dT%H:%M:%S",
- "%Y-%m-%d %H:%M:%S",
- ]
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict["formats"] = self.formats
- return info_dict
-
- def get_metavar(self, param: "Parameter") -> str:
- return f"[{'|'.join(self.formats)}]"
-
- def _try_to_convert_date(self, value: t.Any, format: str) -> t.Optional[datetime]:
- try:
- return datetime.strptime(value, format)
- except ValueError:
- return None
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- if isinstance(value, datetime):
- return value
-
- for format in self.formats:
- converted = self._try_to_convert_date(value, format)
-
- if converted is not None:
- return converted
-
- formats_str = ", ".join(map(repr, self.formats))
- self.fail(
- ngettext(
- "{value!r} does not match the format {format}.",
- "{value!r} does not match the formats {formats}.",
- len(self.formats),
- ).format(value=value, format=formats_str, formats=formats_str),
- param,
- ctx,
- )
-
- def __repr__(self) -> str:
- return "DateTime"
-
-
-class _NumberParamTypeBase(ParamType):
- _number_class: t.ClassVar[t.Type[t.Any]]
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- try:
- return self._number_class(value)
- except ValueError:
- self.fail(
- _("{value!r} is not a valid {number_type}.").format(
- value=value, number_type=self.name
- ),
- param,
- ctx,
- )
-
-
-class _NumberRangeBase(_NumberParamTypeBase):
- def __init__(
- self,
- min: t.Optional[float] = None,
- max: t.Optional[float] = None,
- min_open: bool = False,
- max_open: bool = False,
- clamp: bool = False,
- ) -> None:
- self.min = min
- self.max = max
- self.min_open = min_open
- self.max_open = max_open
- self.clamp = clamp
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict.update(
- min=self.min,
- max=self.max,
- min_open=self.min_open,
- max_open=self.max_open,
- clamp=self.clamp,
- )
- return info_dict
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- import operator
-
- rv = super().convert(value, param, ctx)
- lt_min: bool = self.min is not None and (
- operator.le if self.min_open else operator.lt
- )(rv, self.min)
- gt_max: bool = self.max is not None and (
- operator.ge if self.max_open else operator.gt
- )(rv, self.max)
-
- if self.clamp:
- if lt_min:
- return self._clamp(self.min, 1, self.min_open) # type: ignore
-
- if gt_max:
- return self._clamp(self.max, -1, self.max_open) # type: ignore
-
- if lt_min or gt_max:
- self.fail(
- _("{value} is not in the range {range}.").format(
- value=rv, range=self._describe_range()
- ),
- param,
- ctx,
- )
-
- return rv
-
- def _clamp(self, bound: float, dir: "te.Literal[1, -1]", open: bool) -> float:
- """Find the valid value to clamp to bound in the given
- direction.
-
- :param bound: The boundary value.
- :param dir: 1 or -1 indicating the direction to move.
- :param open: If true, the range does not include the bound.
- """
- raise NotImplementedError
-
- def _describe_range(self) -> str:
- """Describe the range for use in help text."""
- if self.min is None:
- op = "<" if self.max_open else "<="
- return f"x{op}{self.max}"
-
- if self.max is None:
- op = ">" if self.min_open else ">="
- return f"x{op}{self.min}"
-
- lop = "<" if self.min_open else "<="
- rop = "<" if self.max_open else "<="
- return f"{self.min}{lop}x{rop}{self.max}"
-
- def __repr__(self) -> str:
- clamp = " clamped" if self.clamp else ""
- return f"<{type(self).__name__} {self._describe_range()}{clamp}>"
-
-
-class IntParamType(_NumberParamTypeBase):
- name = "integer"
- _number_class = int
-
- def __repr__(self) -> str:
- return "INT"
-
-
-class IntRange(_NumberRangeBase, IntParamType):
- """Restrict an :data:`click.INT` value to a range of accepted
- values. See :ref:`ranges`.
-
- If ``min`` or ``max`` are not passed, any value is accepted in that
- direction. If ``min_open`` or ``max_open`` are enabled, the
- corresponding boundary is not included in the range.
-
- If ``clamp`` is enabled, a value outside the range is clamped to the
- boundary instead of failing.
-
- .. versionchanged:: 8.0
- Added the ``min_open`` and ``max_open`` parameters.
- """
-
- name = "integer range"
-
- def _clamp( # type: ignore
- self, bound: int, dir: "te.Literal[1, -1]", open: bool
- ) -> int:
- if not open:
- return bound
-
- return bound + dir
-
-
-class FloatParamType(_NumberParamTypeBase):
- name = "float"
- _number_class = float
-
- def __repr__(self) -> str:
- return "FLOAT"
-
-
-class FloatRange(_NumberRangeBase, FloatParamType):
- """Restrict a :data:`click.FLOAT` value to a range of accepted
- values. See :ref:`ranges`.
-
- If ``min`` or ``max`` are not passed, any value is accepted in that
- direction. If ``min_open`` or ``max_open`` are enabled, the
- corresponding boundary is not included in the range.
-
- If ``clamp`` is enabled, a value outside the range is clamped to the
- boundary instead of failing. This is not supported if either
- boundary is marked ``open``.
-
- .. versionchanged:: 8.0
- Added the ``min_open`` and ``max_open`` parameters.
- """
-
- name = "float range"
-
- def __init__(
- self,
- min: t.Optional[float] = None,
- max: t.Optional[float] = None,
- min_open: bool = False,
- max_open: bool = False,
- clamp: bool = False,
- ) -> None:
- super().__init__(
- min=min, max=max, min_open=min_open, max_open=max_open, clamp=clamp
- )
-
- if (min_open or max_open) and clamp:
- raise TypeError("Clamping is not supported for open bounds.")
-
- def _clamp(self, bound: float, dir: "te.Literal[1, -1]", open: bool) -> float:
- if not open:
- return bound
-
- # Could use Python 3.9's math.nextafter here, but clamping an
- # open float range doesn't seem to be particularly useful. It's
- # left up to the user to write a callback to do it if needed.
- raise RuntimeError("Clamping is not supported for open bounds.")
-
-
-class BoolParamType(ParamType):
- name = "boolean"
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- if value in {False, True}:
- return bool(value)
-
- norm = value.strip().lower()
-
- if norm in {"1", "true", "t", "yes", "y", "on"}:
- return True
-
- if norm in {"0", "false", "f", "no", "n", "off"}:
- return False
-
- self.fail(
- _("{value!r} is not a valid boolean.").format(value=value), param, ctx
- )
-
- def __repr__(self) -> str:
- return "BOOL"
-
-
-class UUIDParameterType(ParamType):
- name = "uuid"
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- import uuid
-
- if isinstance(value, uuid.UUID):
- return value
-
- value = value.strip()
-
- try:
- return uuid.UUID(value)
- except ValueError:
- self.fail(
- _("{value!r} is not a valid UUID.").format(value=value), param, ctx
- )
-
- def __repr__(self) -> str:
- return "UUID"
-
-
-class File(ParamType):
- """Declares a parameter to be a file for reading or writing. The file
- is automatically closed once the context tears down (after the command
- finished working).
-
- Files can be opened for reading or writing. The special value ``-``
- indicates stdin or stdout depending on the mode.
-
- By default, the file is opened for reading text data, but it can also be
- opened in binary mode or for writing. The encoding parameter can be used
- to force a specific encoding.
-
- The `lazy` flag controls if the file should be opened immediately or upon
- first IO. The default is to be non-lazy for standard input and output
- streams as well as files opened for reading, `lazy` otherwise. When opening a
- file lazily for reading, it is still opened temporarily for validation, but
- will not be held open until first IO. lazy is mainly useful when opening
- for writing to avoid creating the file until it is needed.
-
- Starting with Click 2.0, files can also be opened atomically in which
- case all writes go into a separate file in the same folder and upon
- completion the file will be moved over to the original location. This
- is useful if a file regularly read by other users is modified.
-
- See :ref:`file-args` for more information.
- """
-
- name = "filename"
- envvar_list_splitter: t.ClassVar[str] = os.path.pathsep
-
- def __init__(
- self,
- mode: str = "r",
- encoding: t.Optional[str] = None,
- errors: t.Optional[str] = "strict",
- lazy: t.Optional[bool] = None,
- atomic: bool = False,
- ) -> None:
- self.mode = mode
- self.encoding = encoding
- self.errors = errors
- self.lazy = lazy
- self.atomic = atomic
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict.update(mode=self.mode, encoding=self.encoding)
- return info_dict
-
- def resolve_lazy_flag(self, value: "t.Union[str, os.PathLike[str]]") -> bool:
- if self.lazy is not None:
- return self.lazy
- if os.fspath(value) == "-":
- return False
- elif "w" in self.mode:
- return True
- return False
-
- def convert(
- self,
- value: t.Union[str, "os.PathLike[str]", t.IO[t.Any]],
- param: t.Optional["Parameter"],
- ctx: t.Optional["Context"],
- ) -> t.IO[t.Any]:
- if _is_file_like(value):
- return value
-
- value = t.cast("t.Union[str, os.PathLike[str]]", value)
-
- try:
- lazy = self.resolve_lazy_flag(value)
-
- if lazy:
- lf = LazyFile(
- value, self.mode, self.encoding, self.errors, atomic=self.atomic
- )
-
- if ctx is not None:
- ctx.call_on_close(lf.close_intelligently)
-
- return t.cast(t.IO[t.Any], lf)
-
- f, should_close = open_stream(
- value, self.mode, self.encoding, self.errors, atomic=self.atomic
- )
-
- # If a context is provided, we automatically close the file
- # at the end of the context execution (or flush out). If a
- # context does not exist, it's the caller's responsibility to
- # properly close the file. This for instance happens when the
- # type is used with prompts.
- if ctx is not None:
- if should_close:
- ctx.call_on_close(safecall(f.close))
- else:
- ctx.call_on_close(safecall(f.flush))
-
- return f
- except OSError as e: # noqa: B014
- self.fail(f"'{format_filename(value)}': {e.strerror}", param, ctx)
-
- def shell_complete(
- self, ctx: "Context", param: "Parameter", incomplete: str
- ) -> t.List["CompletionItem"]:
- """Return a special completion marker that tells the completion
- system to use the shell to provide file path completions.
-
- :param ctx: Invocation context for this command.
- :param param: The parameter that is requesting completion.
- :param incomplete: Value being completed. May be empty.
-
- .. versionadded:: 8.0
- """
- from click.shell_completion import CompletionItem
-
- return [CompletionItem(incomplete, type="file")]
-
-
-def _is_file_like(value: t.Any) -> "te.TypeGuard[t.IO[t.Any]]":
- return hasattr(value, "read") or hasattr(value, "write")
-
-
-class Path(ParamType):
- """The ``Path`` type is similar to the :class:`File` type, but
- returns the filename instead of an open file. Various checks can be
- enabled to validate the type of file and permissions.
-
- :param exists: The file or directory needs to exist for the value to
- be valid. If this is not set to ``True``, and the file does not
- exist, then all further checks are silently skipped.
- :param file_okay: Allow a file as a value.
- :param dir_okay: Allow a directory as a value.
- :param readable: if true, a readable check is performed.
- :param writable: if true, a writable check is performed.
- :param executable: if true, an executable check is performed.
- :param resolve_path: Make the value absolute and resolve any
- symlinks. A ``~`` is not expanded, as this is supposed to be
- done by the shell only.
- :param allow_dash: Allow a single dash as a value, which indicates
- a standard stream (but does not open it). Use
- :func:`~click.open_file` to handle opening this value.
- :param path_type: Convert the incoming path value to this type. If
- ``None``, keep Python's default, which is ``str``. Useful to
- convert to :class:`pathlib.Path`.
-
- .. versionchanged:: 8.1
- Added the ``executable`` parameter.
-
- .. versionchanged:: 8.0
- Allow passing ``path_type=pathlib.Path``.
-
- .. versionchanged:: 6.0
- Added the ``allow_dash`` parameter.
- """
-
- envvar_list_splitter: t.ClassVar[str] = os.path.pathsep
-
- def __init__(
- self,
- exists: bool = False,
- file_okay: bool = True,
- dir_okay: bool = True,
- writable: bool = False,
- readable: bool = True,
- resolve_path: bool = False,
- allow_dash: bool = False,
- path_type: t.Optional[t.Type[t.Any]] = None,
- executable: bool = False,
- ):
- self.exists = exists
- self.file_okay = file_okay
- self.dir_okay = dir_okay
- self.readable = readable
- self.writable = writable
- self.executable = executable
- self.resolve_path = resolve_path
- self.allow_dash = allow_dash
- self.type = path_type
-
- if self.file_okay and not self.dir_okay:
- self.name: str = _("file")
- elif self.dir_okay and not self.file_okay:
- self.name = _("directory")
- else:
- self.name = _("path")
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict.update(
- exists=self.exists,
- file_okay=self.file_okay,
- dir_okay=self.dir_okay,
- writable=self.writable,
- readable=self.readable,
- allow_dash=self.allow_dash,
- )
- return info_dict
-
- def coerce_path_result(
- self, value: "t.Union[str, os.PathLike[str]]"
- ) -> "t.Union[str, bytes, os.PathLike[str]]":
- if self.type is not None and not isinstance(value, self.type):
- if self.type is str:
- return os.fsdecode(value)
- elif self.type is bytes:
- return os.fsencode(value)
- else:
- return t.cast("os.PathLike[str]", self.type(value))
-
- return value
-
- def convert(
- self,
- value: "t.Union[str, os.PathLike[str]]",
- param: t.Optional["Parameter"],
- ctx: t.Optional["Context"],
- ) -> "t.Union[str, bytes, os.PathLike[str]]":
- rv = value
-
- is_dash = self.file_okay and self.allow_dash and rv in (b"-", "-")
-
- if not is_dash:
- if self.resolve_path:
- # os.path.realpath doesn't resolve symlinks on Windows
- # until Python 3.8. Use pathlib for now.
- import pathlib
-
- rv = os.fsdecode(pathlib.Path(rv).resolve())
-
- try:
- st = os.stat(rv)
- except OSError:
- if not self.exists:
- return self.coerce_path_result(rv)
- self.fail(
- _("{name} {filename!r} does not exist.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
-
- if not self.file_okay and stat.S_ISREG(st.st_mode):
- self.fail(
- _("{name} {filename!r} is a file.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
- if not self.dir_okay and stat.S_ISDIR(st.st_mode):
- self.fail(
- _("{name} '{filename}' is a directory.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
-
- if self.readable and not os.access(rv, os.R_OK):
- self.fail(
- _("{name} {filename!r} is not readable.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
-
- if self.writable and not os.access(rv, os.W_OK):
- self.fail(
- _("{name} {filename!r} is not writable.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
-
- if self.executable and not os.access(value, os.X_OK):
- self.fail(
- _("{name} {filename!r} is not executable.").format(
- name=self.name.title(), filename=format_filename(value)
- ),
- param,
- ctx,
- )
-
- return self.coerce_path_result(rv)
-
- def shell_complete(
- self, ctx: "Context", param: "Parameter", incomplete: str
- ) -> t.List["CompletionItem"]:
- """Return a special completion marker that tells the completion
- system to use the shell to provide path completions for only
- directories or any paths.
-
- :param ctx: Invocation context for this command.
- :param param: The parameter that is requesting completion.
- :param incomplete: Value being completed. May be empty.
-
- .. versionadded:: 8.0
- """
- from click.shell_completion import CompletionItem
-
- type = "dir" if self.dir_okay and not self.file_okay else "file"
- return [CompletionItem(incomplete, type=type)]
-
-
-class Tuple(CompositeParamType):
- """The default behavior of Click is to apply a type on a value directly.
- This works well in most cases, except for when `nargs` is set to a fixed
- count and different types should be used for different items. In this
- case the :class:`Tuple` type can be used. This type can only be used
- if `nargs` is set to a fixed number.
-
- For more information see :ref:`tuple-type`.
-
- This can be selected by using a Python tuple literal as a type.
-
- :param types: a list of types that should be used for the tuple items.
- """
-
- def __init__(self, types: t.Sequence[t.Union[t.Type[t.Any], ParamType]]) -> None:
- self.types: t.Sequence[ParamType] = [convert_type(ty) for ty in types]
-
- def to_info_dict(self) -> t.Dict[str, t.Any]:
- info_dict = super().to_info_dict()
- info_dict["types"] = [t.to_info_dict() for t in self.types]
- return info_dict
-
- @property
- def name(self) -> str: # type: ignore
- return f"<{' '.join(ty.name for ty in self.types)}>"
-
- @property
- def arity(self) -> int: # type: ignore
- return len(self.types)
-
- def convert(
- self, value: t.Any, param: t.Optional["Parameter"], ctx: t.Optional["Context"]
- ) -> t.Any:
- len_type = len(self.types)
- len_value = len(value)
-
- if len_value != len_type:
- self.fail(
- ngettext(
- "{len_type} values are required, but {len_value} was given.",
- "{len_type} values are required, but {len_value} were given.",
- len_value,
- ).format(len_type=len_type, len_value=len_value),
- param=param,
- ctx=ctx,
- )
-
- return tuple(ty(x, param, ctx) for ty, x in zip(self.types, value))
-
-
-def convert_type(ty: t.Optional[t.Any], default: t.Optional[t.Any] = None) -> ParamType:
- """Find the most appropriate :class:`ParamType` for the given Python
- type. If the type isn't provided, it can be inferred from a default
- value.
- """
- guessed_type = False
-
- if ty is None and default is not None:
- if isinstance(default, (tuple, list)):
- # If the default is empty, ty will remain None and will
- # return STRING.
- if default:
- item = default[0]
-
- # A tuple of tuples needs to detect the inner types.
- # Can't call convert recursively because that would
- # incorrectly unwind the tuple to a single type.
- if isinstance(item, (tuple, list)):
- ty = tuple(map(type, item))
- else:
- ty = type(item)
- else:
- ty = type(default)
-
- guessed_type = True
-
- if isinstance(ty, tuple):
- return Tuple(ty)
-
- if isinstance(ty, ParamType):
- return ty
-
- if ty is str or ty is None:
- return STRING
-
- if ty is int:
- return INT
-
- if ty is float:
- return FLOAT
-
- if ty is bool:
- return BOOL
-
- if guessed_type:
- return STRING
-
- if __debug__:
- try:
- if issubclass(ty, ParamType):
- raise AssertionError(
- f"Attempted to use an uninstantiated parameter type ({ty})."
- )
- except TypeError:
- # ty is an instance (correct), so issubclass fails.
- pass
-
- return FuncParamType(ty)
-
-
-#: A dummy parameter type that just does nothing. From a user's
-#: perspective this appears to just be the same as `STRING` but
-#: internally no string conversion takes place if the input was bytes.
-#: This is usually useful when working with file paths as they can
-#: appear in bytes and unicode.
-#:
-#: For path related uses the :class:`Path` type is a better choice but
-#: there are situations where an unprocessed type is useful which is why
-#: it is is provided.
-#:
-#: .. versionadded:: 4.0
-UNPROCESSED = UnprocessedParamType()
-
-#: A unicode string parameter type which is the implicit default. This
-#: can also be selected by using ``str`` as type.
-STRING = StringParamType()
-
-#: An integer parameter. This can also be selected by using ``int`` as
-#: type.
-INT = IntParamType()
-
-#: A floating point value parameter. This can also be selected by using
-#: ``float`` as type.
-FLOAT = FloatParamType()
-
-#: A boolean parameter. This is the default for boolean flags. This can
-#: also be selected by using ``bool`` as a type.
-BOOL = BoolParamType()
-
-#: A UUID parameter.
-UUID = UUIDParameterType()
diff --git a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/h11/tests/test_connection.py b/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/h11/tests/test_connection.py
deleted file mode 100644
index 73a27b98bebd949cb3b99e19a3a8a484455b58d7..0000000000000000000000000000000000000000
--- a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/h11/tests/test_connection.py
+++ /dev/null
@@ -1,1122 +0,0 @@
-from typing import Any, cast, Dict, List, Optional, Tuple, Type
-
-import pytest
-
-from .._connection import _body_framing, _keep_alive, Connection, NEED_DATA, PAUSED
-from .._events import (
- ConnectionClosed,
- Data,
- EndOfMessage,
- Event,
- InformationalResponse,
- Request,
- Response,
-)
-from .._state import (
- CLIENT,
- CLOSED,
- DONE,
- ERROR,
- IDLE,
- MIGHT_SWITCH_PROTOCOL,
- MUST_CLOSE,
- SEND_BODY,
- SEND_RESPONSE,
- SERVER,
- SWITCHED_PROTOCOL,
-)
-from .._util import LocalProtocolError, RemoteProtocolError, Sentinel
-from .helpers import ConnectionPair, get_all_events, receive_and_get
-
-
-def test__keep_alive() -> None:
- assert _keep_alive(
- Request(method="GET", target="/", headers=[("Host", "Example.com")])
- )
- assert not _keep_alive(
- Request(
- method="GET",
- target="/",
- headers=[("Host", "Example.com"), ("Connection", "close")],
- )
- )
- assert not _keep_alive(
- Request(
- method="GET",
- target="/",
- headers=[("Host", "Example.com"), ("Connection", "a, b, cLOse, foo")],
- )
- )
- assert not _keep_alive(
- Request(method="GET", target="/", headers=[], http_version="1.0") # type: ignore[arg-type]
- )
-
- assert _keep_alive(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- assert not _keep_alive(Response(status_code=200, headers=[("Connection", "close")]))
- assert not _keep_alive(
- Response(status_code=200, headers=[("Connection", "a, b, cLOse, foo")])
- )
- assert not _keep_alive(Response(status_code=200, headers=[], http_version="1.0")) # type: ignore[arg-type]
-
-
-def test__body_framing() -> None:
- def headers(cl: Optional[int], te: bool) -> List[Tuple[str, str]]:
- headers = []
- if cl is not None:
- headers.append(("Content-Length", str(cl)))
- if te:
- headers.append(("Transfer-Encoding", "chunked"))
- return headers
-
- def resp(
- status_code: int = 200, cl: Optional[int] = None, te: bool = False
- ) -> Response:
- return Response(status_code=status_code, headers=headers(cl, te))
-
- def req(cl: Optional[int] = None, te: bool = False) -> Request:
- h = headers(cl, te)
- h += [("Host", "example.com")]
- return Request(method="GET", target="/", headers=h)
-
- # Special cases where the headers are ignored:
- for kwargs in [{}, {"cl": 100}, {"te": True}, {"cl": 100, "te": True}]:
- kwargs = cast(Dict[str, Any], kwargs)
- for meth, r in [
- (b"HEAD", resp(**kwargs)),
- (b"GET", resp(status_code=204, **kwargs)),
- (b"GET", resp(status_code=304, **kwargs)),
- ]:
- assert _body_framing(meth, r) == ("content-length", (0,))
-
- # Transfer-encoding
- for kwargs in [{"te": True}, {"cl": 100, "te": True}]:
- kwargs = cast(Dict[str, Any], kwargs)
- for meth, r in [(None, req(**kwargs)), (b"GET", resp(**kwargs))]: # type: ignore
- assert _body_framing(meth, r) == ("chunked", ())
-
- # Content-Length
- for meth, r in [(None, req(cl=100)), (b"GET", resp(cl=100))]: # type: ignore
- assert _body_framing(meth, r) == ("content-length", (100,))
-
- # No headers
- assert _body_framing(None, req()) == ("content-length", (0,)) # type: ignore
- assert _body_framing(b"GET", resp()) == ("http/1.0", ())
-
-
-def test_Connection_basics_and_content_length() -> None:
- with pytest.raises(ValueError):
- Connection("CLIENT") # type: ignore
-
- p = ConnectionPair()
- assert p.conn[CLIENT].our_role is CLIENT
- assert p.conn[CLIENT].their_role is SERVER
- assert p.conn[SERVER].our_role is SERVER
- assert p.conn[SERVER].their_role is CLIENT
-
- data = p.send(
- CLIENT,
- Request(
- method="GET",
- target="/",
- headers=[("Host", "example.com"), ("Content-Length", "10")],
- ),
- )
- assert data == (
- b"GET / HTTP/1.1\r\n" b"Host: example.com\r\n" b"Content-Length: 10\r\n\r\n"
- )
-
- for conn in p.conns:
- assert conn.states == {CLIENT: SEND_BODY, SERVER: SEND_RESPONSE}
- assert p.conn[CLIENT].our_state is SEND_BODY
- assert p.conn[CLIENT].their_state is SEND_RESPONSE
- assert p.conn[SERVER].our_state is SEND_RESPONSE
- assert p.conn[SERVER].their_state is SEND_BODY
-
- assert p.conn[CLIENT].their_http_version is None
- assert p.conn[SERVER].their_http_version == b"1.1"
-
- data = p.send(SERVER, InformationalResponse(status_code=100, headers=[])) # type: ignore[arg-type]
- assert data == b"HTTP/1.1 100 \r\n\r\n"
-
- data = p.send(SERVER, Response(status_code=200, headers=[("Content-Length", "11")]))
- assert data == b"HTTP/1.1 200 \r\nContent-Length: 11\r\n\r\n"
-
- for conn in p.conns:
- assert conn.states == {CLIENT: SEND_BODY, SERVER: SEND_BODY}
-
- assert p.conn[CLIENT].their_http_version == b"1.1"
- assert p.conn[SERVER].their_http_version == b"1.1"
-
- data = p.send(CLIENT, Data(data=b"12345"))
- assert data == b"12345"
- data = p.send(
- CLIENT, Data(data=b"67890"), expect=[Data(data=b"67890"), EndOfMessage()]
- )
- assert data == b"67890"
- data = p.send(CLIENT, EndOfMessage(), expect=[])
- assert data == b""
-
- for conn in p.conns:
- assert conn.states == {CLIENT: DONE, SERVER: SEND_BODY}
-
- data = p.send(SERVER, Data(data=b"1234567890"))
- assert data == b"1234567890"
- data = p.send(SERVER, Data(data=b"1"), expect=[Data(data=b"1"), EndOfMessage()])
- assert data == b"1"
- data = p.send(SERVER, EndOfMessage(), expect=[])
- assert data == b""
-
- for conn in p.conns:
- assert conn.states == {CLIENT: DONE, SERVER: DONE}
-
-
-def test_chunked() -> None:
- p = ConnectionPair()
-
- p.send(
- CLIENT,
- Request(
- method="GET",
- target="/",
- headers=[("Host", "example.com"), ("Transfer-Encoding", "chunked")],
- ),
- )
- data = p.send(CLIENT, Data(data=b"1234567890", chunk_start=True, chunk_end=True))
- assert data == b"a\r\n1234567890\r\n"
- data = p.send(CLIENT, Data(data=b"abcde", chunk_start=True, chunk_end=True))
- assert data == b"5\r\nabcde\r\n"
- data = p.send(CLIENT, Data(data=b""), expect=[])
- assert data == b""
- data = p.send(CLIENT, EndOfMessage(headers=[("hello", "there")]))
- assert data == b"0\r\nhello: there\r\n\r\n"
-
- p.send(
- SERVER, Response(status_code=200, headers=[("Transfer-Encoding", "chunked")])
- )
- p.send(SERVER, Data(data=b"54321", chunk_start=True, chunk_end=True))
- p.send(SERVER, Data(data=b"12345", chunk_start=True, chunk_end=True))
- p.send(SERVER, EndOfMessage())
-
- for conn in p.conns:
- assert conn.states == {CLIENT: DONE, SERVER: DONE}
-
-
-def test_chunk_boundaries() -> None:
- conn = Connection(our_role=SERVER)
-
- request = (
- b"POST / HTTP/1.1\r\n"
- b"Host: example.com\r\n"
- b"Transfer-Encoding: chunked\r\n"
- b"\r\n"
- )
- conn.receive_data(request)
- assert conn.next_event() == Request(
- method="POST",
- target="/",
- headers=[("Host", "example.com"), ("Transfer-Encoding", "chunked")],
- )
- assert conn.next_event() is NEED_DATA
-
- conn.receive_data(b"5\r\nhello\r\n")
- assert conn.next_event() == Data(data=b"hello", chunk_start=True, chunk_end=True)
-
- conn.receive_data(b"5\r\nhel")
- assert conn.next_event() == Data(data=b"hel", chunk_start=True, chunk_end=False)
-
- conn.receive_data(b"l")
- assert conn.next_event() == Data(data=b"l", chunk_start=False, chunk_end=False)
-
- conn.receive_data(b"o\r\n")
- assert conn.next_event() == Data(data=b"o", chunk_start=False, chunk_end=True)
-
- conn.receive_data(b"5\r\nhello")
- assert conn.next_event() == Data(data=b"hello", chunk_start=True, chunk_end=True)
-
- conn.receive_data(b"\r\n")
- assert conn.next_event() == NEED_DATA
-
- conn.receive_data(b"0\r\n\r\n")
- assert conn.next_event() == EndOfMessage()
-
-
-def test_client_talking_to_http10_server() -> None:
- c = Connection(CLIENT)
- c.send(Request(method="GET", target="/", headers=[("Host", "example.com")]))
- c.send(EndOfMessage())
- assert c.our_state is DONE
- # No content-length, so Http10 framing for body
- assert receive_and_get(c, b"HTTP/1.0 200 OK\r\n\r\n") == [
- Response(status_code=200, headers=[], http_version="1.0", reason=b"OK") # type: ignore[arg-type]
- ]
- assert c.our_state is MUST_CLOSE
- assert receive_and_get(c, b"12345") == [Data(data=b"12345")]
- assert receive_and_get(c, b"67890") == [Data(data=b"67890")]
- assert receive_and_get(c, b"") == [EndOfMessage(), ConnectionClosed()]
- assert c.their_state is CLOSED
-
-
-def test_server_talking_to_http10_client() -> None:
- c = Connection(SERVER)
- # No content-length, so no body
- # NB: no host header
- assert receive_and_get(c, b"GET / HTTP/1.0\r\n\r\n") == [
- Request(method="GET", target="/", headers=[], http_version="1.0"), # type: ignore[arg-type]
- EndOfMessage(),
- ]
- assert c.their_state is MUST_CLOSE
-
- # We automatically Connection: close back at them
- assert (
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- == b"HTTP/1.1 200 \r\nConnection: close\r\n\r\n"
- )
-
- assert c.send(Data(data=b"12345")) == b"12345"
- assert c.send(EndOfMessage()) == b""
- assert c.our_state is MUST_CLOSE
-
- # Check that it works if they do send Content-Length
- c = Connection(SERVER)
- # NB: no host header
- assert receive_and_get(c, b"POST / HTTP/1.0\r\nContent-Length: 10\r\n\r\n1") == [
- Request(
- method="POST",
- target="/",
- headers=[("Content-Length", "10")],
- http_version="1.0",
- ),
- Data(data=b"1"),
- ]
- assert receive_and_get(c, b"234567890") == [Data(data=b"234567890"), EndOfMessage()]
- assert c.their_state is MUST_CLOSE
- assert receive_and_get(c, b"") == [ConnectionClosed()]
-
-
-def test_automatic_transfer_encoding_in_response() -> None:
- # Check that in responses, the user can specify either Transfer-Encoding:
- # chunked or no framing at all, and in both cases we automatically select
- # the right option depending on whether the peer speaks HTTP/1.0 or
- # HTTP/1.1
- for user_headers in [
- [("Transfer-Encoding", "chunked")],
- [],
- # In fact, this even works if Content-Length is set,
- # because if both are set then Transfer-Encoding wins
- [("Transfer-Encoding", "chunked"), ("Content-Length", "100")],
- ]:
- user_headers = cast(List[Tuple[str, str]], user_headers)
- p = ConnectionPair()
- p.send(
- CLIENT,
- [
- Request(method="GET", target="/", headers=[("Host", "example.com")]),
- EndOfMessage(),
- ],
- )
- # When speaking to HTTP/1.1 client, all of the above cases get
- # normalized to Transfer-Encoding: chunked
- p.send(
- SERVER,
- Response(status_code=200, headers=user_headers),
- expect=Response(
- status_code=200, headers=[("Transfer-Encoding", "chunked")]
- ),
- )
-
- # When speaking to HTTP/1.0 client, all of the above cases get
- # normalized to no-framing-headers
- c = Connection(SERVER)
- receive_and_get(c, b"GET / HTTP/1.0\r\n\r\n")
- assert (
- c.send(Response(status_code=200, headers=user_headers))
- == b"HTTP/1.1 200 \r\nConnection: close\r\n\r\n"
- )
- assert c.send(Data(data=b"12345")) == b"12345"
-
-
-def test_automagic_connection_close_handling() -> None:
- p = ConnectionPair()
- # If the user explicitly sets Connection: close, then we notice and
- # respect it
- p.send(
- CLIENT,
- [
- Request(
- method="GET",
- target="/",
- headers=[("Host", "example.com"), ("Connection", "close")],
- ),
- EndOfMessage(),
- ],
- )
- for conn in p.conns:
- assert conn.states[CLIENT] is MUST_CLOSE
- # And if the client sets it, the server automatically echoes it back
- p.send(
- SERVER,
- # no header here...
- [Response(status_code=204, headers=[]), EndOfMessage()], # type: ignore[arg-type]
- # ...but oh look, it arrived anyway
- expect=[
- Response(status_code=204, headers=[("connection", "close")]),
- EndOfMessage(),
- ],
- )
- for conn in p.conns:
- assert conn.states == {CLIENT: MUST_CLOSE, SERVER: MUST_CLOSE}
-
-
-def test_100_continue() -> None:
- def setup() -> ConnectionPair:
- p = ConnectionPair()
- p.send(
- CLIENT,
- Request(
- method="GET",
- target="/",
- headers=[
- ("Host", "example.com"),
- ("Content-Length", "100"),
- ("Expect", "100-continue"),
- ],
- ),
- )
- for conn in p.conns:
- assert conn.client_is_waiting_for_100_continue
- assert not p.conn[CLIENT].they_are_waiting_for_100_continue
- assert p.conn[SERVER].they_are_waiting_for_100_continue
- return p
-
- # Disabled by 100 Continue
- p = setup()
- p.send(SERVER, InformationalResponse(status_code=100, headers=[])) # type: ignore[arg-type]
- for conn in p.conns:
- assert not conn.client_is_waiting_for_100_continue
- assert not conn.they_are_waiting_for_100_continue
-
- # Disabled by a real response
- p = setup()
- p.send(
- SERVER, Response(status_code=200, headers=[("Transfer-Encoding", "chunked")])
- )
- for conn in p.conns:
- assert not conn.client_is_waiting_for_100_continue
- assert not conn.they_are_waiting_for_100_continue
-
- # Disabled by the client going ahead and sending stuff anyway
- p = setup()
- p.send(CLIENT, Data(data=b"12345"))
- for conn in p.conns:
- assert not conn.client_is_waiting_for_100_continue
- assert not conn.they_are_waiting_for_100_continue
-
-
-def test_max_incomplete_event_size_countermeasure() -> None:
- # Infinitely long headers are definitely not okay
- c = Connection(SERVER)
- c.receive_data(b"GET / HTTP/1.0\r\nEndless: ")
- assert c.next_event() is NEED_DATA
- with pytest.raises(RemoteProtocolError):
- while True:
- c.receive_data(b"a" * 1024)
- c.next_event()
-
- # Checking that the same header is accepted / rejected depending on the
- # max_incomplete_event_size setting:
- c = Connection(SERVER, max_incomplete_event_size=5000)
- c.receive_data(b"GET / HTTP/1.0\r\nBig: ")
- c.receive_data(b"a" * 4000)
- c.receive_data(b"\r\n\r\n")
- assert get_all_events(c) == [
- Request(
- method="GET", target="/", http_version="1.0", headers=[("big", "a" * 4000)]
- ),
- EndOfMessage(),
- ]
-
- c = Connection(SERVER, max_incomplete_event_size=4000)
- c.receive_data(b"GET / HTTP/1.0\r\nBig: ")
- c.receive_data(b"a" * 4000)
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
- # Temporarily exceeding the size limit is fine, as long as its done with
- # complete events:
- c = Connection(SERVER, max_incomplete_event_size=5000)
- c.receive_data(b"GET / HTTP/1.0\r\nContent-Length: 10000")
- c.receive_data(b"\r\n\r\n" + b"a" * 10000)
- assert get_all_events(c) == [
- Request(
- method="GET",
- target="/",
- http_version="1.0",
- headers=[("Content-Length", "10000")],
- ),
- Data(data=b"a" * 10000),
- EndOfMessage(),
- ]
-
- c = Connection(SERVER, max_incomplete_event_size=100)
- # Two pipelined requests to create a way-too-big receive buffer... but
- # it's fine because we're not checking
- c.receive_data(
- b"GET /1 HTTP/1.1\r\nHost: a\r\n\r\n"
- b"GET /2 HTTP/1.1\r\nHost: b\r\n\r\n" + b"X" * 1000
- )
- assert get_all_events(c) == [
- Request(method="GET", target="/1", headers=[("host", "a")]),
- EndOfMessage(),
- ]
- # Even more data comes in, still no problem
- c.receive_data(b"X" * 1000)
- # We can respond and reuse to get the second pipelined request
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- c.start_next_cycle()
- assert get_all_events(c) == [
- Request(method="GET", target="/2", headers=[("host", "b")]),
- EndOfMessage(),
- ]
- # But once we unpause and try to read the next message, and find that it's
- # incomplete and the buffer is *still* way too large, then *that's* a
- # problem:
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- c.start_next_cycle()
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-def test_reuse_simple() -> None:
- p = ConnectionPair()
- p.send(
- CLIENT,
- [Request(method="GET", target="/", headers=[("Host", "a")]), EndOfMessage()],
- )
- p.send(
- SERVER,
- [
- Response(status_code=200, headers=[(b"transfer-encoding", b"chunked")]),
- EndOfMessage(),
- ],
- )
- for conn in p.conns:
- assert conn.states == {CLIENT: DONE, SERVER: DONE}
- conn.start_next_cycle()
-
- p.send(
- CLIENT,
- [
- Request(method="DELETE", target="/foo", headers=[("Host", "a")]),
- EndOfMessage(),
- ],
- )
- p.send(
- SERVER,
- [
- Response(status_code=404, headers=[(b"transfer-encoding", b"chunked")]),
- EndOfMessage(),
- ],
- )
-
-
-def test_pipelining() -> None:
- # Client doesn't support pipelining, so we have to do this by hand
- c = Connection(SERVER)
- assert c.next_event() is NEED_DATA
- # 3 requests all bunched up
- c.receive_data(
- b"GET /1 HTTP/1.1\r\nHost: a.com\r\nContent-Length: 5\r\n\r\n"
- b"12345"
- b"GET /2 HTTP/1.1\r\nHost: a.com\r\nContent-Length: 5\r\n\r\n"
- b"67890"
- b"GET /3 HTTP/1.1\r\nHost: a.com\r\n\r\n"
- )
- assert get_all_events(c) == [
- Request(
- method="GET",
- target="/1",
- headers=[("Host", "a.com"), ("Content-Length", "5")],
- ),
- Data(data=b"12345"),
- EndOfMessage(),
- ]
- assert c.their_state is DONE
- assert c.our_state is SEND_RESPONSE
-
- assert c.next_event() is PAUSED
-
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- assert c.their_state is DONE
- assert c.our_state is DONE
-
- c.start_next_cycle()
-
- assert get_all_events(c) == [
- Request(
- method="GET",
- target="/2",
- headers=[("Host", "a.com"), ("Content-Length", "5")],
- ),
- Data(data=b"67890"),
- EndOfMessage(),
- ]
- assert c.next_event() is PAUSED
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- c.start_next_cycle()
-
- assert get_all_events(c) == [
- Request(method="GET", target="/3", headers=[("Host", "a.com")]),
- EndOfMessage(),
- ]
- # Doesn't pause this time, no trailing data
- assert c.next_event() is NEED_DATA
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
-
- # Arrival of more data triggers pause
- assert c.next_event() is NEED_DATA
- c.receive_data(b"SADF")
- assert c.next_event() is PAUSED
- assert c.trailing_data == (b"SADF", False)
- # If EOF arrives while paused, we don't see that either:
- c.receive_data(b"")
- assert c.trailing_data == (b"SADF", True)
- assert c.next_event() is PAUSED
- c.receive_data(b"")
- assert c.next_event() is PAUSED
- # Can't call receive_data with non-empty buf after closing it
- with pytest.raises(RuntimeError):
- c.receive_data(b"FDSA")
-
-
-def test_protocol_switch() -> None:
- for (req, deny, accept) in [
- (
- Request(
- method="CONNECT",
- target="example.com:443",
- headers=[("Host", "foo"), ("Content-Length", "1")],
- ),
- Response(status_code=404, headers=[(b"transfer-encoding", b"chunked")]),
- Response(status_code=200, headers=[(b"transfer-encoding", b"chunked")]),
- ),
- (
- Request(
- method="GET",
- target="/",
- headers=[("Host", "foo"), ("Content-Length", "1"), ("Upgrade", "a, b")],
- ),
- Response(status_code=200, headers=[(b"transfer-encoding", b"chunked")]),
- InformationalResponse(status_code=101, headers=[("Upgrade", "a")]),
- ),
- (
- Request(
- method="CONNECT",
- target="example.com:443",
- headers=[("Host", "foo"), ("Content-Length", "1"), ("Upgrade", "a, b")],
- ),
- Response(status_code=404, headers=[(b"transfer-encoding", b"chunked")]),
- # Accept CONNECT, not upgrade
- Response(status_code=200, headers=[(b"transfer-encoding", b"chunked")]),
- ),
- (
- Request(
- method="CONNECT",
- target="example.com:443",
- headers=[("Host", "foo"), ("Content-Length", "1"), ("Upgrade", "a, b")],
- ),
- Response(status_code=404, headers=[(b"transfer-encoding", b"chunked")]),
- # Accept Upgrade, not CONNECT
- InformationalResponse(status_code=101, headers=[("Upgrade", "b")]),
- ),
- ]:
-
- def setup() -> ConnectionPair:
- p = ConnectionPair()
- p.send(CLIENT, req)
- # No switch-related state change stuff yet; the client has to
- # finish the request before that kicks in
- for conn in p.conns:
- assert conn.states[CLIENT] is SEND_BODY
- p.send(CLIENT, [Data(data=b"1"), EndOfMessage()])
- for conn in p.conns:
- assert conn.states[CLIENT] is MIGHT_SWITCH_PROTOCOL
- assert p.conn[SERVER].next_event() is PAUSED
- return p
-
- # Test deny case
- p = setup()
- p.send(SERVER, deny)
- for conn in p.conns:
- assert conn.states == {CLIENT: DONE, SERVER: SEND_BODY}
- p.send(SERVER, EndOfMessage())
- # Check that re-use is still allowed after a denial
- for conn in p.conns:
- conn.start_next_cycle()
-
- # Test accept case
- p = setup()
- p.send(SERVER, accept)
- for conn in p.conns:
- assert conn.states == {CLIENT: SWITCHED_PROTOCOL, SERVER: SWITCHED_PROTOCOL}
- conn.receive_data(b"123")
- assert conn.next_event() is PAUSED
- conn.receive_data(b"456")
- assert conn.next_event() is PAUSED
- assert conn.trailing_data == (b"123456", False)
-
- # Pausing in might-switch, then recovery
- # (weird artificial case where the trailing data actually is valid
- # HTTP for some reason, because this makes it easier to test the state
- # logic)
- p = setup()
- sc = p.conn[SERVER]
- sc.receive_data(b"GET / HTTP/1.0\r\n\r\n")
- assert sc.next_event() is PAUSED
- assert sc.trailing_data == (b"GET / HTTP/1.0\r\n\r\n", False)
- sc.send(deny)
- assert sc.next_event() is PAUSED
- sc.send(EndOfMessage())
- sc.start_next_cycle()
- assert get_all_events(sc) == [
- Request(method="GET", target="/", headers=[], http_version="1.0"), # type: ignore[arg-type]
- EndOfMessage(),
- ]
-
- # When we're DONE, have no trailing data, and the connection gets
- # closed, we report ConnectionClosed(). When we're in might-switch or
- # switched, we don't.
- p = setup()
- sc = p.conn[SERVER]
- sc.receive_data(b"")
- assert sc.next_event() is PAUSED
- assert sc.trailing_data == (b"", True)
- p.send(SERVER, accept)
- assert sc.next_event() is PAUSED
-
- p = setup()
- sc = p.conn[SERVER]
- sc.receive_data(b"")
- assert sc.next_event() is PAUSED
- sc.send(deny)
- assert sc.next_event() == ConnectionClosed()
-
- # You can't send after switching protocols, or while waiting for a
- # protocol switch
- p = setup()
- with pytest.raises(LocalProtocolError):
- p.conn[CLIENT].send(
- Request(method="GET", target="/", headers=[("Host", "a")])
- )
- p = setup()
- p.send(SERVER, accept)
- with pytest.raises(LocalProtocolError):
- p.conn[SERVER].send(Data(data=b"123"))
-
-
-def test_close_simple() -> None:
- # Just immediately closing a new connection without anything having
- # happened yet.
- for (who_shot_first, who_shot_second) in [(CLIENT, SERVER), (SERVER, CLIENT)]:
-
- def setup() -> ConnectionPair:
- p = ConnectionPair()
- p.send(who_shot_first, ConnectionClosed())
- for conn in p.conns:
- assert conn.states == {
- who_shot_first: CLOSED,
- who_shot_second: MUST_CLOSE,
- }
- return p
-
- # You can keep putting b"" into a closed connection, and you keep
- # getting ConnectionClosed() out:
- p = setup()
- assert p.conn[who_shot_second].next_event() == ConnectionClosed()
- assert p.conn[who_shot_second].next_event() == ConnectionClosed()
- p.conn[who_shot_second].receive_data(b"")
- assert p.conn[who_shot_second].next_event() == ConnectionClosed()
- # Second party can close...
- p = setup()
- p.send(who_shot_second, ConnectionClosed())
- for conn in p.conns:
- assert conn.our_state is CLOSED
- assert conn.their_state is CLOSED
- # But trying to receive new data on a closed connection is a
- # RuntimeError (not ProtocolError, because the problem here isn't
- # violation of HTTP, it's violation of physics)
- p = setup()
- with pytest.raises(RuntimeError):
- p.conn[who_shot_second].receive_data(b"123")
- # And receiving new data on a MUST_CLOSE connection is a ProtocolError
- p = setup()
- p.conn[who_shot_first].receive_data(b"GET")
- with pytest.raises(RemoteProtocolError):
- p.conn[who_shot_first].next_event()
-
-
-def test_close_different_states() -> None:
- req = [
- Request(method="GET", target="/foo", headers=[("Host", "a")]),
- EndOfMessage(),
- ]
- resp = [
- Response(status_code=200, headers=[(b"transfer-encoding", b"chunked")]),
- EndOfMessage(),
- ]
-
- # Client before request
- p = ConnectionPair()
- p.send(CLIENT, ConnectionClosed())
- for conn in p.conns:
- assert conn.states == {CLIENT: CLOSED, SERVER: MUST_CLOSE}
-
- # Client after request
- p = ConnectionPair()
- p.send(CLIENT, req)
- p.send(CLIENT, ConnectionClosed())
- for conn in p.conns:
- assert conn.states == {CLIENT: CLOSED, SERVER: SEND_RESPONSE}
-
- # Server after request -> not allowed
- p = ConnectionPair()
- p.send(CLIENT, req)
- with pytest.raises(LocalProtocolError):
- p.conn[SERVER].send(ConnectionClosed())
- p.conn[CLIENT].receive_data(b"")
- with pytest.raises(RemoteProtocolError):
- p.conn[CLIENT].next_event()
-
- # Server after response
- p = ConnectionPair()
- p.send(CLIENT, req)
- p.send(SERVER, resp)
- p.send(SERVER, ConnectionClosed())
- for conn in p.conns:
- assert conn.states == {CLIENT: MUST_CLOSE, SERVER: CLOSED}
-
- # Both after closing (ConnectionClosed() is idempotent)
- p = ConnectionPair()
- p.send(CLIENT, req)
- p.send(SERVER, resp)
- p.send(CLIENT, ConnectionClosed())
- p.send(SERVER, ConnectionClosed())
- p.send(CLIENT, ConnectionClosed())
- p.send(SERVER, ConnectionClosed())
-
- # In the middle of sending -> not allowed
- p = ConnectionPair()
- p.send(
- CLIENT,
- Request(
- method="GET", target="/", headers=[("Host", "a"), ("Content-Length", "10")]
- ),
- )
- with pytest.raises(LocalProtocolError):
- p.conn[CLIENT].send(ConnectionClosed())
- p.conn[SERVER].receive_data(b"")
- with pytest.raises(RemoteProtocolError):
- p.conn[SERVER].next_event()
-
-
-# Receive several requests and then client shuts down their side of the
-# connection; we can respond to each
-def test_pipelined_close() -> None:
- c = Connection(SERVER)
- # 2 requests then a close
- c.receive_data(
- b"GET /1 HTTP/1.1\r\nHost: a.com\r\nContent-Length: 5\r\n\r\n"
- b"12345"
- b"GET /2 HTTP/1.1\r\nHost: a.com\r\nContent-Length: 5\r\n\r\n"
- b"67890"
- )
- c.receive_data(b"")
- assert get_all_events(c) == [
- Request(
- method="GET",
- target="/1",
- headers=[("host", "a.com"), ("content-length", "5")],
- ),
- Data(data=b"12345"),
- EndOfMessage(),
- ]
- assert c.states[CLIENT] is DONE
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- assert c.states[SERVER] is DONE
- c.start_next_cycle()
- assert get_all_events(c) == [
- Request(
- method="GET",
- target="/2",
- headers=[("host", "a.com"), ("content-length", "5")],
- ),
- Data(data=b"67890"),
- EndOfMessage(),
- ConnectionClosed(),
- ]
- assert c.states == {CLIENT: CLOSED, SERVER: SEND_RESPONSE}
- c.send(Response(status_code=200, headers=[])) # type: ignore[arg-type]
- c.send(EndOfMessage())
- assert c.states == {CLIENT: CLOSED, SERVER: MUST_CLOSE}
- c.send(ConnectionClosed())
- assert c.states == {CLIENT: CLOSED, SERVER: CLOSED}
-
-
-def test_sendfile() -> None:
- class SendfilePlaceholder:
- def __len__(self) -> int:
- return 10
-
- placeholder = SendfilePlaceholder()
-
- def setup(
- header: Tuple[str, str], http_version: str
- ) -> Tuple[Connection, Optional[List[bytes]]]:
- c = Connection(SERVER)
- receive_and_get(
- c, "GET / HTTP/{}\r\nHost: a\r\n\r\n".format(http_version).encode("ascii")
- )
- headers = []
- if header:
- headers.append(header)
- c.send(Response(status_code=200, headers=headers))
- return c, c.send_with_data_passthrough(Data(data=placeholder)) # type: ignore
-
- c, data = setup(("Content-Length", "10"), "1.1")
- assert data == [placeholder] # type: ignore
- # Raises an error if the connection object doesn't think we've sent
- # exactly 10 bytes
- c.send(EndOfMessage())
-
- _, data = setup(("Transfer-Encoding", "chunked"), "1.1")
- assert placeholder in data # type: ignore
- data[data.index(placeholder)] = b"x" * 10 # type: ignore
- assert b"".join(data) == b"a\r\nxxxxxxxxxx\r\n" # type: ignore
-
- c, data = setup(None, "1.0") # type: ignore
- assert data == [placeholder] # type: ignore
- assert c.our_state is SEND_BODY
-
-
-def test_errors() -> None:
- # After a receive error, you can't receive
- for role in [CLIENT, SERVER]:
- c = Connection(our_role=role)
- c.receive_data(b"gibberish\r\n\r\n")
- with pytest.raises(RemoteProtocolError):
- c.next_event()
- # Now any attempt to receive continues to raise
- assert c.their_state is ERROR
- assert c.our_state is not ERROR
- print(c._cstate.states)
- with pytest.raises(RemoteProtocolError):
- c.next_event()
- # But we can still yell at the client for sending us gibberish
- if role is SERVER:
- assert (
- c.send(Response(status_code=400, headers=[])) # type: ignore[arg-type]
- == b"HTTP/1.1 400 \r\nConnection: close\r\n\r\n"
- )
-
- # After an error sending, you can no longer send
- # (This is especially important for things like content-length errors,
- # where there's complex internal state being modified)
- def conn(role: Type[Sentinel]) -> Connection:
- c = Connection(our_role=role)
- if role is SERVER:
- # Put it into the state where it *could* send a response...
- receive_and_get(c, b"GET / HTTP/1.0\r\n\r\n")
- assert c.our_state is SEND_RESPONSE
- return c
-
- for role in [CLIENT, SERVER]:
- if role is CLIENT:
- # This HTTP/1.0 request won't be detected as bad until after we go
- # through the state machine and hit the writing code
- good = Request(method="GET", target="/", headers=[("Host", "example.com")])
- bad = Request(
- method="GET",
- target="/",
- headers=[("Host", "example.com")],
- http_version="1.0",
- )
- elif role is SERVER:
- good = Response(status_code=200, headers=[]) # type: ignore[arg-type,assignment]
- bad = Response(status_code=200, headers=[], http_version="1.0") # type: ignore[arg-type,assignment]
- # Make sure 'good' actually is good
- c = conn(role)
- c.send(good)
- assert c.our_state is not ERROR
- # Do that again, but this time sending 'bad' first
- c = conn(role)
- with pytest.raises(LocalProtocolError):
- c.send(bad)
- assert c.our_state is ERROR
- assert c.their_state is not ERROR
- # Now 'good' is not so good
- with pytest.raises(LocalProtocolError):
- c.send(good)
-
- # And check send_failed() too
- c = conn(role)
- c.send_failed()
- assert c.our_state is ERROR
- assert c.their_state is not ERROR
- # This is idempotent
- c.send_failed()
- assert c.our_state is ERROR
- assert c.their_state is not ERROR
-
-
-def test_idle_receive_nothing() -> None:
- # At one point this incorrectly raised an error
- for role in [CLIENT, SERVER]:
- c = Connection(role)
- assert c.next_event() is NEED_DATA
-
-
-def test_connection_drop() -> None:
- c = Connection(SERVER)
- c.receive_data(b"GET /")
- assert c.next_event() is NEED_DATA
- c.receive_data(b"")
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-def test_408_request_timeout() -> None:
- # Should be able to send this spontaneously as a server without seeing
- # anything from client
- p = ConnectionPair()
- p.send(SERVER, Response(status_code=408, headers=[(b"connection", b"close")]))
-
-
-# This used to raise IndexError
-def test_empty_request() -> None:
- c = Connection(SERVER)
- c.receive_data(b"\r\n")
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-# This used to raise IndexError
-def test_empty_response() -> None:
- c = Connection(CLIENT)
- c.send(Request(method="GET", target="/", headers=[("Host", "a")]))
- c.receive_data(b"\r\n")
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-@pytest.mark.parametrize(
- "data",
- [
- b"\x00",
- b"\x20",
- b"\x16\x03\x01\x00\xa5", # Typical start of a TLS Client Hello
- ],
-)
-def test_early_detection_of_invalid_request(data: bytes) -> None:
- c = Connection(SERVER)
- # Early detection should occur before even receiving a `\r\n`
- c.receive_data(data)
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-@pytest.mark.parametrize(
- "data",
- [
- b"\x00",
- b"\x20",
- b"\x16\x03\x03\x00\x31", # Typical start of a TLS Server Hello
- ],
-)
-def test_early_detection_of_invalid_response(data: bytes) -> None:
- c = Connection(CLIENT)
- # Early detection should occur before even receiving a `\r\n`
- c.receive_data(data)
- with pytest.raises(RemoteProtocolError):
- c.next_event()
-
-
-# This used to give different headers for HEAD and GET.
-# The correct way to handle HEAD is to put whatever headers we *would* have
-# put if it were a GET -- even though we know that for HEAD, those headers
-# will be ignored.
-def test_HEAD_framing_headers() -> None:
- def setup(method: bytes, http_version: bytes) -> Connection:
- c = Connection(SERVER)
- c.receive_data(
- method + b" / HTTP/" + http_version + b"\r\n" + b"Host: example.com\r\n\r\n"
- )
- assert type(c.next_event()) is Request
- assert type(c.next_event()) is EndOfMessage
- return c
-
- for method in [b"GET", b"HEAD"]:
- # No Content-Length, HTTP/1.1 peer, should use chunked
- c = setup(method, b"1.1")
- assert (
- c.send(Response(status_code=200, headers=[])) == b"HTTP/1.1 200 \r\n" # type: ignore[arg-type]
- b"Transfer-Encoding: chunked\r\n\r\n"
- )
-
- # No Content-Length, HTTP/1.0 peer, frame with connection: close
- c = setup(method, b"1.0")
- assert (
- c.send(Response(status_code=200, headers=[])) == b"HTTP/1.1 200 \r\n" # type: ignore[arg-type]
- b"Connection: close\r\n\r\n"
- )
-
- # Content-Length + Transfer-Encoding, TE wins
- c = setup(method, b"1.1")
- assert (
- c.send(
- Response(
- status_code=200,
- headers=[
- ("Content-Length", "100"),
- ("Transfer-Encoding", "chunked"),
- ],
- )
- )
- == b"HTTP/1.1 200 \r\n"
- b"Transfer-Encoding: chunked\r\n\r\n"
- )
-
-
-def test_special_exceptions_for_lost_connection_in_message_body() -> None:
- c = Connection(SERVER)
- c.receive_data(
- b"POST / HTTP/1.1\r\n" b"Host: example.com\r\n" b"Content-Length: 100\r\n\r\n"
- )
- assert type(c.next_event()) is Request
- assert c.next_event() is NEED_DATA
- c.receive_data(b"12345")
- assert c.next_event() == Data(data=b"12345")
- c.receive_data(b"")
- with pytest.raises(RemoteProtocolError) as excinfo:
- c.next_event()
- assert "received 5 bytes" in str(excinfo.value)
- assert "expected 100" in str(excinfo.value)
-
- c = Connection(SERVER)
- c.receive_data(
- b"POST / HTTP/1.1\r\n"
- b"Host: example.com\r\n"
- b"Transfer-Encoding: chunked\r\n\r\n"
- )
- assert type(c.next_event()) is Request
- assert c.next_event() is NEED_DATA
- c.receive_data(b"8\r\n012345")
- assert c.next_event().data == b"012345" # type: ignore
- c.receive_data(b"")
- with pytest.raises(RemoteProtocolError) as excinfo:
- c.next_event()
- assert "incomplete chunked read" in str(excinfo.value)
diff --git a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/httpx/_content.py b/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/httpx/_content.py
deleted file mode 100644
index b16e12d954327e7ecd5f05885bb8778a0fbfa047..0000000000000000000000000000000000000000
--- a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/httpx/_content.py
+++ /dev/null
@@ -1,238 +0,0 @@
-import inspect
-import warnings
-from json import dumps as json_dumps
-from typing import (
- Any,
- AsyncIterable,
- AsyncIterator,
- Dict,
- Iterable,
- Iterator,
- Mapping,
- Optional,
- Tuple,
- Union,
-)
-from urllib.parse import urlencode
-
-from ._exceptions import StreamClosed, StreamConsumed
-from ._multipart import MultipartStream
-from ._types import (
- AsyncByteStream,
- RequestContent,
- RequestData,
- RequestFiles,
- ResponseContent,
- SyncByteStream,
-)
-from ._utils import peek_filelike_length, primitive_value_to_str
-
-
-class ByteStream(AsyncByteStream, SyncByteStream):
- def __init__(self, stream: bytes) -> None:
- self._stream = stream
-
- def __iter__(self) -> Iterator[bytes]:
- yield self._stream
-
- async def __aiter__(self) -> AsyncIterator[bytes]:
- yield self._stream
-
-
-class IteratorByteStream(SyncByteStream):
- CHUNK_SIZE = 65_536
-
- def __init__(self, stream: Iterable[bytes]):
- self._stream = stream
- self._is_stream_consumed = False
- self._is_generator = inspect.isgenerator(stream)
-
- def __iter__(self) -> Iterator[bytes]:
- if self._is_stream_consumed and self._is_generator:
- raise StreamConsumed()
-
- self._is_stream_consumed = True
- if hasattr(self._stream, "read"):
- # File-like interfaces should use 'read' directly.
- chunk = self._stream.read(self.CHUNK_SIZE)
- while chunk:
- yield chunk
- chunk = self._stream.read(self.CHUNK_SIZE)
- else:
- # Otherwise iterate.
- for part in self._stream:
- yield part
-
-
-class AsyncIteratorByteStream(AsyncByteStream):
- CHUNK_SIZE = 65_536
-
- def __init__(self, stream: AsyncIterable[bytes]):
- self._stream = stream
- self._is_stream_consumed = False
- self._is_generator = inspect.isasyncgen(stream)
-
- async def __aiter__(self) -> AsyncIterator[bytes]:
- if self._is_stream_consumed and self._is_generator:
- raise StreamConsumed()
-
- self._is_stream_consumed = True
- if hasattr(self._stream, "aread"):
- # File-like interfaces should use 'aread' directly.
- chunk = await self._stream.aread(self.CHUNK_SIZE)
- while chunk:
- yield chunk
- chunk = await self._stream.aread(self.CHUNK_SIZE)
- else:
- # Otherwise iterate.
- async for part in self._stream:
- yield part
-
-
-class UnattachedStream(AsyncByteStream, SyncByteStream):
- """
- If a request or response is serialized using pickle, then it is no longer
- attached to a stream for I/O purposes. Any stream operations should result
- in `httpx.StreamClosed`.
- """
-
- def __iter__(self) -> Iterator[bytes]:
- raise StreamClosed()
-
- async def __aiter__(self) -> AsyncIterator[bytes]:
- raise StreamClosed()
- yield b"" # pragma: no cover
-
-
-def encode_content(
- content: Union[str, bytes, Iterable[bytes], AsyncIterable[bytes]]
-) -> Tuple[Dict[str, str], Union[SyncByteStream, AsyncByteStream]]:
- if isinstance(content, (bytes, str)):
- body = content.encode("utf-8") if isinstance(content, str) else content
- content_length = len(body)
- headers = {"Content-Length": str(content_length)} if body else {}
- return headers, ByteStream(body)
-
- elif isinstance(content, Iterable) and not isinstance(content, dict):
- # `not isinstance(content, dict)` is a bit oddly specific, but it
- # catches a case that's easy for users to make in error, and would
- # otherwise pass through here, like any other bytes-iterable,
- # because `dict` happens to be iterable. See issue #2491.
- content_length_or_none = peek_filelike_length(content)
-
- if content_length_or_none is None:
- headers = {"Transfer-Encoding": "chunked"}
- else:
- headers = {"Content-Length": str(content_length_or_none)}
- return headers, IteratorByteStream(content) # type: ignore
-
- elif isinstance(content, AsyncIterable):
- headers = {"Transfer-Encoding": "chunked"}
- return headers, AsyncIteratorByteStream(content)
-
- raise TypeError(f"Unexpected type for 'content', {type(content)!r}")
-
-
-def encode_urlencoded_data(
- data: RequestData,
-) -> Tuple[Dict[str, str], ByteStream]:
- plain_data = []
- for key, value in data.items():
- if isinstance(value, (list, tuple)):
- plain_data.extend([(key, primitive_value_to_str(item)) for item in value])
- else:
- plain_data.append((key, primitive_value_to_str(value)))
- body = urlencode(plain_data, doseq=True).encode("utf-8")
- content_length = str(len(body))
- content_type = "application/x-www-form-urlencoded"
- headers = {"Content-Length": content_length, "Content-Type": content_type}
- return headers, ByteStream(body)
-
-
-def encode_multipart_data(
- data: RequestData, files: RequestFiles, boundary: Optional[bytes]
-) -> Tuple[Dict[str, str], MultipartStream]:
- multipart = MultipartStream(data=data, files=files, boundary=boundary)
- headers = multipart.get_headers()
- return headers, multipart
-
-
-def encode_text(text: str) -> Tuple[Dict[str, str], ByteStream]:
- body = text.encode("utf-8")
- content_length = str(len(body))
- content_type = "text/plain; charset=utf-8"
- headers = {"Content-Length": content_length, "Content-Type": content_type}
- return headers, ByteStream(body)
-
-
-def encode_html(html: str) -> Tuple[Dict[str, str], ByteStream]:
- body = html.encode("utf-8")
- content_length = str(len(body))
- content_type = "text/html; charset=utf-8"
- headers = {"Content-Length": content_length, "Content-Type": content_type}
- return headers, ByteStream(body)
-
-
-def encode_json(json: Any) -> Tuple[Dict[str, str], ByteStream]:
- body = json_dumps(json).encode("utf-8")
- content_length = str(len(body))
- content_type = "application/json"
- headers = {"Content-Length": content_length, "Content-Type": content_type}
- return headers, ByteStream(body)
-
-
-def encode_request(
- content: Optional[RequestContent] = None,
- data: Optional[RequestData] = None,
- files: Optional[RequestFiles] = None,
- json: Optional[Any] = None,
- boundary: Optional[bytes] = None,
-) -> Tuple[Dict[str, str], Union[SyncByteStream, AsyncByteStream]]:
- """
- Handles encoding the given `content`, `data`, `files`, and `json`,
- returning a two-tuple of (, ).
- """
- if data is not None and not isinstance(data, Mapping):
- # We prefer to separate `content=`
- # for raw request content, and `data=