title
stringlengths
1
300
score
int64
0
8.54k
selftext
stringlengths
0
40k
created
timestamp[ns]date
2023-04-01 04:30:41
2025-06-30 03:16:29
url
stringlengths
0
878
author
stringlengths
3
20
domain
stringlengths
0
82
edited
timestamp[ns]date
1970-01-01 00:00:00
2025-06-26 17:30:18
gilded
int64
0
2
gildings
stringclasses
7 values
id
stringlengths
7
7
locked
bool
2 classes
media
stringlengths
646
1.8k
name
stringlengths
10
10
permalink
stringlengths
33
82
spoiler
bool
2 classes
stickied
bool
2 classes
thumbnail
stringlengths
4
213
ups
int64
0
8.54k
preview
stringlengths
301
5.01k
Someone should Abliterate the R1 models/help me figure out how to do it.
0
title. this could be interesting. I have 128gb of cpu ram if someone could give me a rundown on how to do it I would be glad to if cpu ram can even do that.
2025-01-21T03:25:55
https://www.reddit.com/r/LocalLLaMA/comments/1i69dds/someone_should_abliterate_the_r1_modelshelp_me/
ElectricalAngle1611
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i69dds
false
null
t3_1i69dds
/r/LocalLLaMA/comments/1i69dds/someone_should_abliterate_the_r1_modelshelp_me/
false
false
self
0
null
Deepseek R1 (Ollama) Hardware benchmark for LocalLLM
147
Deepseek R1 was released and looks like one of the best models for local LLM. I tested it on some GPUs to see how many tps it can achieve. Tests were run on Ollama. Input prompt: How to {build a pc|build a website|build xxx}? Thoughts: \- \`deepseek-r1:14b\` can run on any GPU without a significant performance gap. \- \`deepseek-r1:32b\` runs better on a single GPU with \~24GB VRAM: RTX 3090 offers the best price/performance. RTX Titan is acceptable. \- \`deepseek-r1:70b\` performs best with 2 x RTX 3090 (17tps) in terms of price/performance. However, it doubles the electricity cost compared to RTX 6000 ADA (19tps) or RTX A6000 (12tps). \- \`M3 Max 40GPU\` has high memory but only delivers 3-7 tps for \`deepseek-r1:70b\`. It is also loud, and the GPU temperature is high (> 90 C).
2025-01-21T03:26:07
https://www.reddit.com/r/LocalLLaMA/comments/1i69dhz/deepseek_r1_ollama_hardware_benchmark_for_localllm/
Joehua87
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i69dhz
false
null
t3_1i69dhz
/r/LocalLLaMA/comments/1i69dhz/deepseek_r1_ollama_hardware_benchmark_for_localllm/
false
false
self
147
null
Quen2.5-Coder-32B-Instruct-FP16 + 4x AMD Instinct Mi60 Server
2
2025-01-21T04:11:29
https://v.redd.it/98r4yjkcv9ee1
Any_Praline_8178
/r/LocalLLaMA/comments/1i6a7pq/quen25coder32binstructfp16_4x_amd_instinct_mi60/
1970-01-01T00:00:00
0
{}
1i6a7pq
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/98r4yjkcv9ee1/DASHPlaylist.mpd?a=1740154295%2CZTYzYzI0ZDE3MmVhNDQ0NTQ3YmE1MDM3YzNjZjVhOWFmNjMxZDRiZjljM2QyZWUxMmMzMDI2ZDEwMGUyYWQzZQ%3D%3D&v=1&f=sd', 'duration': 184, 'fallback_url': 'https://v.redd.it/98r4yjkcv9ee1/DASH_1080.mp4?source=fallback', 'has_audio': False, 'height': 1904, 'hls_url': 'https://v.redd.it/98r4yjkcv9ee1/HLSPlaylist.m3u8?a=1740154295%2CYzk5YjA5NzcxOWUzZGU2YmU0ZGZmZTY5MTEwNzE5MjdjNmMzMTBiMWZhOGIyN2MxZDNhMjIyMmZmNzEzNjU0ZA%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/98r4yjkcv9ee1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 1080}}
t3_1i6a7pq
/r/LocalLLaMA/comments/1i6a7pq/quen25coder32binstructfp16_4x_amd_instinct_mi60/
false
false
https://external-preview…5aa4cd56829d7807
2
{'enabled': False, 'images': [{'id': 'Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35', 'resolutions': [{'height': 190, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=108&crop=smart&format=pjpg&auto=webp&s=c2749932c0b50b73c2c4daccb3288bb5722ddcc0', 'width': 108}, {'height': 380, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=216&crop=smart&format=pjpg&auto=webp&s=3d2ec02f6201cdd57f7453e3b5ae4e004b0a3e2a', 'width': 216}, {'height': 563, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=320&crop=smart&format=pjpg&auto=webp&s=2baedfd0d7d20d61af4995a40a06539777db64ea', 'width': 320}, {'height': 1127, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=640&crop=smart&format=pjpg&auto=webp&s=9f310b3cea2b411994fd1f3d3c597f84c4109e29', 'width': 640}, {'height': 1691, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=960&crop=smart&format=pjpg&auto=webp&s=17f7f1f58e5a4a52ec9b2b8b7de4bc39ac3a96b7', 'width': 960}, {'height': 1903, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?width=1080&crop=smart&format=pjpg&auto=webp&s=8ba656b545c046886b1b8a3b4f2c28deea0f87d3', 'width': 1080}], 'source': {'height': 3796, 'url': 'https://external-preview.redd.it/Z21wcnFqa2N2OWVlMSf761CrKXe2i5qyFnzLhNsPSHWiJDWU06Z_Yc1Rfn35.png?format=pjpg&auto=webp&s=5ba7d94ec239b67e0b047927bd097652e1358624', 'width': 2154}, 'variants': {}}]}
deepseek r1 gets there eventually
13
>>> The surgeon, who is the boy's father says, "I can't operate on this boy, he's my son!" Who is the surgeon to the boy? <think> Okay, so I came across this riddle that goes like this: "The surgeon, who is the boy's father says, 'I can't operate on this boy, he's my son!' Who is the surgeon to the boy?" At first glance, it seems a bit confusing because of how it's phrased. Let me try to break it down step by step. First, let's look at the structure of the sentence. The main part is "The surgeon says, 'I can't operate on this boy, he's my son!'" So, the surgeon is making a statement about why they can't perform the surgery: because the boy is their son. But wait, that seems contradictory because if the surgeon is the father of the boy, then naturally, the surgeon would be the boy's father, which makes sense in a straightforward way. But riddles often play on words or assumptions we make without thinking deeply. So maybe there's a twist here. Let me think about possible angles. One common angle in such riddles is to consider that "surgeon" could refer to someone else besides the boy's father. For instance, sometimes these puzzles involve roles where a woman might be the surgeon instead of a man, challenging the stereotype that surgeons are typically male. So, maybe the surgeon is actually the boy's mother. Wait, but in this case, the surgeon says, "he's my son," which implies that the surgeon has a son who is the patient. If the surgeon were the mother, she could still say, "he's my son." So, perhaps the riddle is trying to make us think beyond the immediate assumption of the father being the surgeon. Alternatively, maybe it's about the relationship in terms of roles rather than gender. But I'm not sure if that's the case here. Let me try to rephrase the sentence for clarity: "The surgeon (who is the boy's father) says, 'I can't operate on this boy; he's my son!'" So, it directly states that the surgeon is the boy's father. Therefore, the answer should be straightforward: the surgeon is the boy's father. But then why is this presented as a riddle? Maybe because people assume that surgeons are male, but perhaps the intended answer is that the surgeon is the mother instead. However, the sentence explicitly states that the surgeon is the boy's father, so that might not be it. Wait, could there be another layer? Maybe the surgeon isn't referring to the boy as their son in a familial sense but in some other context, like being a student or an apprentice. But that seems less likely because the statement is made by the surgeon in a professional setting, talking about operating on a patient. Alternatively, perhaps it's a play on words where "my son" doesn't mean biological son but someone else, like a stepson or adopted child. However, the riddle doesn't provide enough information to suggest that. Let me think again: The structure is "The surgeon, who is the boy's father..." So it clearly states that the surgeon is the father of the boy. Therefore, when the surgeon says, "he's my son," it reinforces that relationship. So logically, the answer is that the surgeon is the boy's father. But maybe the riddle is trying to make us question our assumptions about gender roles in professions. If someone assumes surgeons are male and thus concludes the father is the surgeon, but perhaps the answer is intended to be the mother because she can also be a surgeon. But since the sentence explicitly states that the surgeon is the boy's father, that might not hold. Alternatively, could it be that the surgeon is the boy's parent regardless of gender? So the answer is simply that the surgeon is the boy's parent, but the riddle specifies "father," so that wouldn't make sense. Wait, perhaps the confusion comes from the way it's phrased. Let me parse the sentence again: "The surgeon, who is the boy's father says, 'I can't operate on this boy; he's my son!'" So, the essential information is that the surgeon is the boy's father and that he cannot perform surgery because the boy is his son. So, putting it all together, the surgeon is the boy's father. There doesn't seem to be any trick or twist beyond that. The riddle might be trying to lead us into overcomplicating it by suggesting a different relationship, but the information given directly points to the surgeon being the father. </think> The surgeon is the boy's father.
2025-01-21T04:25:33
https://www.reddit.com/r/LocalLLaMA/comments/1i6agzw/deepseek_r1_gets_there_eventually/
dairypharmer
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6agzw
false
null
t3_1i6agzw
/r/LocalLLaMA/comments/1i6agzw/deepseek_r1_gets_there_eventually/
false
false
self
13
null
The Chinese OBLITERATED OpenAI. A side-by-side comparison of DeepSeek R1 vs OpenAI O1 for Finance
1
[removed]
2025-01-21T04:41:53
https://nexustrade.io/blog/the-chinese-obliterated-openai-a-side-by-side-comparison-of-deepseek-r1-vs-openai-o1-for-finance-20250121
No-Definition-2886
nexustrade.io
1970-01-01T00:00:00
0
{}
1i6ar8y
false
null
t3_1i6ar8y
/r/LocalLLaMA/comments/1i6ar8y/the_chinese_obliterated_openai_a_sidebyside/
false
false
https://b.thumbs.redditm…gHCiWUTNI2vg.jpg
1
{'enabled': False, 'images': [{'id': '5-gqoFCUgqojYgYG-6P580Nt744fEwLXDo8N-3FsIpo', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/y_F5U-sV65j8RddIJ0MewcxY_NOPraI7B85wWVFgPqg.jpg?width=108&crop=smart&auto=webp&s=12fafb6377ef3112cc8e896a7df31d93b2382845', 'width': 108}, {'height': 216, 'url': 'https://external-preview.redd.it/y_F5U-sV65j8RddIJ0MewcxY_NOPraI7B85wWVFgPqg.jpg?width=216&crop=smart&auto=webp&s=8a75d0549844cdc0bb8e661e481ad09f18251ac8', 'width': 216}, {'height': 320, 'url': 'https://external-preview.redd.it/y_F5U-sV65j8RddIJ0MewcxY_NOPraI7B85wWVFgPqg.jpg?width=320&crop=smart&auto=webp&s=8d19cf6543071db9852bee87e93a38d196d84997', 'width': 320}], 'source': {'height': 512, 'url': 'https://external-preview.redd.it/y_F5U-sV65j8RddIJ0MewcxY_NOPraI7B85wWVFgPqg.jpg?auto=webp&s=faf9aff5db2ee4858955c4464ee39b0da1063cac', 'width': 512}, 'variants': {}}]}
AI model and hosting solutions for chatbot
1
[removed]
2025-01-21T04:44:03
https://www.reddit.com/r/LocalLLaMA/comments/1i6aslg/ai_model_and_hosting_solutions_for_chatbot/
No-Salamander6725
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6aslg
false
null
t3_1i6aslg
/r/LocalLLaMA/comments/1i6aslg/ai_model_and_hosting_solutions_for_chatbot/
false
false
self
1
null
I calculated the effective cost of R1 Vs o1 and here's what I found
59
In order to calculate the effective cost of R1 Vs o1, we need to know 2 things: 1. how much each model costs per million output tokens. 2. how much tokens each model generates on average per Chain-of-Thought. You might think: Wait, we can't see o1's CoT since OpenAI hides it, right? While OpenAI does hide the internal CoTs when using o1 via ChatGPT and the API, they did reveal full non-summarized CoTs in the initial announcement of o1-preview ([Source](https://openai.com/index/learning-to-reason-with-llms/)). Later, when o1-2024-1217 was released in December, OpenAI stated, >o1 uses on average 60% fewer reasoning tokens than o1-preview for a given request ([Source](https://openai.com/index/o1-and-new-tools-for-developers/)). Thus, we can calculate the average for o1 by multiplying o1-preview’s token averages by 0.4. The Chain-of-Thought character count per example OpenAI showed us is as follows, as well as the exact same question on R1 below: o1 - \[(16577 + 4475 + 20248 + 12276 + 2930 + 3397 + 2265 + 3542)\*0.4\]/8 = 3285.5 characters per CoT. R1 - (14777 + 14911 + 54837 + 35459 + 7795 + 24143 + 7361 + 4115)/8 = 20424.75 characters per CoT. 20424.75/3285.5 ≈ 6.22 R1 generates 6.22x more reasoning tokens on average than o1 according to the official examples average. R1 costs $2.19/1M output tokens. o1 costs $60/1M output tokens. 60/2.19 ≈ 27.4 o1 costs 27.4x more than R1 price-per-token, however, generates 6.22x fewer tokens. 27.4/6.22 ≈ 4.41 # **therefore in practice R1 is only 4.41x cheaper than o1** (note assumptions made): If o1 generates x less characters it will also be roughly x less tokens. This assumption is fair, however, the precise exact values can vary slightly but should not effect things noticeably. This is just API discussion if you use R1 via the website or the app its infinitely cheaper since its free Vs $20/mo.
2025-01-21T04:52:08
https://www.reddit.com/r/LocalLLaMA/comments/1i6axmv/i_calculated_the_effective_cost_of_r1_vs_o1_and/
pigeon57434
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6axmv
false
null
t3_1i6axmv
/r/LocalLLaMA/comments/1i6axmv/i_calculated_the_effective_cost_of_r1_vs_o1_and/
false
false
self
59
null
how to check for hallucination?
2
Is it possible to train an agent or do existing agents exists to check for hallucinations? what's the best approach for checking for hallucination?
2025-01-21T04:57:24
https://www.reddit.com/r/LocalLLaMA/comments/1i6b0wj/how_to_check_for_hallucination/
Lazy_Wedding_1383
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6b0wj
false
null
t3_1i6b0wj
/r/LocalLLaMA/comments/1i6b0wj/how_to_check_for_hallucination/
false
false
self
2
null
Legal issues around Serpapi and llms
1
Hey guys, I don't know a better place to post this given it covers llm, legal aspects and business. Is it possible to use serp api to query and use it as part of my content for my LLM? Like my llm uses those sources content. This LLM is for my product which I am planning to monetize. More than serapi, I am concerned about copyright issues, let's say news agencies from whom I will use the articles to keep my llm informed. If not serp api, it might be legal to use News API?
2025-01-21T04:58:19
https://www.reddit.com/r/LocalLLaMA/comments/1i6b1hf/legal_issues_around_serpapi_and_llms/
EmergencySherbert247
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6b1hf
false
null
t3_1i6b1hf
/r/LocalLLaMA/comments/1i6b1hf/legal_issues_around_serpapi_and_llms/
false
false
self
1
null
llama3.1 helps me browse and add connections on linkedin, open source
1
2025-01-21T05:02:08
https://youtube.com/shorts/mEzUAmZeAo4
Own-Spell1660
youtube.com
1970-01-01T00:00:00
0
{}
1i6b42g
false
null
t3_1i6b42g
/r/LocalLLaMA/comments/1i6b42g/llama31_helps_me_browse_and_add_connections_on/
false
false
default
1
null
DeepSeek-R1-8B-FP16 + vLLM + 4x AMD Instinct Mi60 Server
9
2025-01-21T05:05:10
https://v.redd.it/1ywwhkmx4aee1
Any_Praline_8178
/r/LocalLLaMA/comments/1i6b5x3/deepseekr18bfp16_vllm_4x_amd_instinct_mi60_server/
1970-01-01T00:00:00
0
{}
1i6b5x3
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/1ywwhkmx4aee1/DASHPlaylist.mpd?a=1740157515%2CYzNhY2NiOWY2NjM1ZmFmNzY4OGIxZDFlZGViODNkMDQwNTczMGExNTQ2MDg4MTMxOWE4MzAxNDBkNzRlOTc2NQ%3D%3D&v=1&f=sd', 'duration': 134, 'fallback_url': 'https://v.redd.it/1ywwhkmx4aee1/DASH_1080.mp4?source=fallback', 'has_audio': False, 'height': 1904, 'hls_url': 'https://v.redd.it/1ywwhkmx4aee1/HLSPlaylist.m3u8?a=1740157515%2CM2M2NjM4MjVmOTBjNzhmMjA3MTk2ZDg5NzJjMzRjNTlmNTk1MzFjYWIxZjcxNGVlYmMwYmE1OTM4YTMxMzczYg%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/1ywwhkmx4aee1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 1080}}
t3_1i6b5x3
/r/LocalLLaMA/comments/1i6b5x3/deepseekr18bfp16_vllm_4x_amd_instinct_mi60_server/
false
false
https://external-preview…490b534d0aaf38d0
9
{'enabled': False, 'images': [{'id': 'Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N', 'resolutions': [{'height': 190, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=108&crop=smart&format=pjpg&auto=webp&s=454e7615a0b237f02480bc6effd4cfdbc02bfbb5', 'width': 108}, {'height': 380, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=216&crop=smart&format=pjpg&auto=webp&s=80329d063c2d50c53cf8a2c685251d9b1ac93e76', 'width': 216}, {'height': 563, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=320&crop=smart&format=pjpg&auto=webp&s=2d2db5b4e32eec76d60d9a8ccd3a7832b5cc72bb', 'width': 320}, {'height': 1127, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=640&crop=smart&format=pjpg&auto=webp&s=539d4d4585f401fc1a9655c886531d1bc66b455a', 'width': 640}, {'height': 1691, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=960&crop=smart&format=pjpg&auto=webp&s=03f4e921a07c396232e42b87846d4b0dbcbb54c8', 'width': 960}, {'height': 1903, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?width=1080&crop=smart&format=pjpg&auto=webp&s=b7b0ecaa3d270c5f6e822123e9b068f1fd60d90c', 'width': 1080}], 'source': {'height': 3796, 'url': 'https://external-preview.redd.it/Yzl1OTZrbXg0YWVlMdktyt9BQ2tR6rOTvvtjjZ1_KkIo_S8J74AmAnZkxV8N.png?format=pjpg&auto=webp&s=70a767a9dc7df1bf7dadefa9d692d1e2b9ef5bf4', 'width': 2154}, 'variants': {}}]}
Better R1 Experience in open webui
131
I just created a simple open webui function for R1 models, it can do the following: 1. Replace the simple <think> tags with <details>& <summary> tags, which makes R1's thoughts collapsible. 2. Remove R1's thoughts in multi-turn conversion, according to deepseeks API docs you should always remove R1's previous thoughts in a multi-turn conversion. Github: [https://github.com/AaronFeng753/Better-R1](https://github.com/AaronFeng753/Better-R1) https://i.redd.it/uqfg9jsy4aee1.gif
2025-01-21T05:05:31
https://www.reddit.com/r/LocalLLaMA/comments/1i6b65q/better_r1_experience_in_open_webui/
AaronFeng47
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6b65q
false
null
t3_1i6b65q
/r/LocalLLaMA/comments/1i6b65q/better_r1_experience_in_open_webui/
false
false
https://b.thumbs.redditm…xOXz6LqSKgDg.jpg
131
{'enabled': False, 'images': [{'id': 'pxsFXTllfI36lsMdRn_i5Hqu9cZErqxRlwTEfV8O2RE', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=108&crop=smart&auto=webp&s=9c0bb6c806f0eedc0b810d9d12af50d1f085be0a', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=216&crop=smart&auto=webp&s=7a2133d6e32fcee96a8fc20786170edda1d52ca5', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=320&crop=smart&auto=webp&s=578fa4972104daf8a9cab37c3b5e2a331744908c', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=640&crop=smart&auto=webp&s=76fb2b05f6cb5884488366b666d5cd74022016bf', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=960&crop=smart&auto=webp&s=883d85bd5407b9f3671e28b6afbfd03864026b1d', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?width=1080&crop=smart&auto=webp&s=bc06dd438d302992db8a8e2dca9212161613d18c', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/YSLEhbNfDODEMmwhWhGpt4WFY87CROwBvvlWm6wYG0k.jpg?auto=webp&s=906f3a2064bd45ba05454a310d4e23925b57427a', 'width': 1200}, 'variants': {}}]}
A fun test for the reasoning models
3
we are going to play a counting game. We will start at 1 and take turns adding to the total. We may add 1 2 or 3 to the total, but you may not increase the total by the same amount as the previous player. The player forced to say a number higher than or equal to 24 loses the game, if you understand start with the number 1. if you need clarification on the rules please ask With two players it's a solved game, but it takes a massive amount of thought to come to the optimal solution. R1 did a pretty good job in two players (won 3 lost 2) but loses consistently at 3 players o1 is hit or miss on two players(won 2 lost 3), and loses consistently at 3 players Gemini lost with two players in my two tests, and loses with 3 players as well.
2025-01-21T05:16:23
https://www.reddit.com/r/LocalLLaMA/comments/1i6bcli/a_fun_test_for_the_reasoning_models/
Mescallan
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6bcli
false
null
t3_1i6bcli
/r/LocalLLaMA/comments/1i6bcli/a_fun_test_for_the_reasoning_models/
false
false
self
3
null
Need Vision Model Recommendations
7
Title. Targeted task is to convert images of handwritten mathematical expressions into Markdown/LaTEX expressions. I have used MiniCPM for other tasks last year but would like to get some updated suggestions.
2025-01-21T05:17:27
https://www.reddit.com/r/LocalLLaMA/comments/1i6bd7z/need_vision_model_recommendations/
Otherwise_Bonus6789
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6bd7z
false
null
t3_1i6bd7z
/r/LocalLLaMA/comments/1i6bd7z/need_vision_model_recommendations/
false
false
self
7
null
Will this be possible with r1?
51
2025-01-21T05:36:23
https://i.redd.it/03nq7sm8aaee1.jpeg
Notdesciplined
i.redd.it
1970-01-01T00:00:00
0
{}
1i6bogs
false
null
t3_1i6bogs
/r/LocalLLaMA/comments/1i6bogs/will_this_be_possible_with_r1/
false
false
https://b.thumbs.redditm…17vg3ysFn6GA.jpg
51
{'enabled': True, 'images': [{'id': 'mBMr6vJkCT2mUBOADwT45BT3IjwS3rBdHhPMEE3LUeU', 'resolutions': [{'height': 67, 'url': 'https://preview.redd.it/03nq7sm8aaee1.jpeg?width=108&crop=smart&auto=webp&s=ade0b1c5f1fb8a3fb5a36fea36b6a5bc8ba9f0e3', 'width': 108}, {'height': 134, 'url': 'https://preview.redd.it/03nq7sm8aaee1.jpeg?width=216&crop=smart&auto=webp&s=5d8e8ddf430bda174ca13ac43ac84d072f5850aa', 'width': 216}, {'height': 199, 'url': 'https://preview.redd.it/03nq7sm8aaee1.jpeg?width=320&crop=smart&auto=webp&s=77f8a7cb588a8cfc1376872ef80b9410ac0186c7', 'width': 320}, {'height': 398, 'url': 'https://preview.redd.it/03nq7sm8aaee1.jpeg?width=640&crop=smart&auto=webp&s=e2edea9a1b2d614a561fa76f2ad39fd61186f448', 'width': 640}], 'source': {'height': 561, 'url': 'https://preview.redd.it/03nq7sm8aaee1.jpeg?auto=webp&s=a7dc31475c569adf66dbfc6ff9066dc23c2bfd94', 'width': 900}, 'variants': {}}]}
You can enable GPU Acceleration in Jan to get better performance thanks to llama.cpp! (Settings -> Advanced Settings -> GPU Acceleration)
1
2025-01-21T05:38:46
https://i.redd.it/xw45sm4xaaee1.png
Embarrassed-Tax-7006
i.redd.it
1970-01-01T00:00:00
0
{}
1i6bpu3
false
null
t3_1i6bpu3
/r/LocalLLaMA/comments/1i6bpu3/you_can_enable_gpu_acceleration_in_jan_to_get/
false
false
https://b.thumbs.redditm…WH6E1e5NBASg.jpg
1
{'enabled': True, 'images': [{'id': 'zWbAoGpOdZ7L_LVxAta_AB7V9zqa8-ojgf8yhhFAcME', 'resolutions': [{'height': 72, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=108&crop=smart&auto=webp&s=9ba5fc1b964ac521da783fc8b9aa12f728bceb21', 'width': 108}, {'height': 145, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=216&crop=smart&auto=webp&s=b0d2010305e9a8e068c0b250b36eb27094ba6b9d', 'width': 216}, {'height': 216, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=320&crop=smart&auto=webp&s=2ba89f7ce158dad86ff8b6938a1cbabf29be717e', 'width': 320}, {'height': 432, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=640&crop=smart&auto=webp&s=7d79e0da39968ccfa56aafc3041b177511e90775', 'width': 640}, {'height': 648, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=960&crop=smart&auto=webp&s=bc08d18b14c76b83d1e8be486283b5436b56d162', 'width': 960}, {'height': 729, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?width=1080&crop=smart&auto=webp&s=067f25fbe931f52bfcdf8422e08bc9d268d12543', 'width': 1080}], 'source': {'height': 1080, 'url': 'https://preview.redd.it/xw45sm4xaaee1.png?auto=webp&s=063fc0ca61619b6d38d4a7e0a6cbac57faf30941', 'width': 1600}, 'variants': {}}]}
What techniques can I try for high quality retrieval on complex PDFs?
4
I have a bunch of tariff PDFs (each PDF talking about different tariffs). And I want to build a RAG system on them. I have tried several different things, but I am still not getting accurate retrieval for certain tariffs - there are some tariffs for which the retrieval is good, but there are other tariffs where the retrieval is not good at all, and the reason I think is because maybe the words used in those texts are not too different (because at the end of the day, these are just definitions and calculations of different kinds of tariffs - so many of the words in them will be similar). And since retrieval depends on texts being dissimilar, it's not going to be good on similar docs. Here are some of the things I've tried: * Approaches: * naive RAG * hybrid RAG * corrective RAG * reflective RAG * agentic RAG (retrieve -> grade docs -> rewrite if retrieval is not good -> retrieve & grade again -> generate if retrieval grading is good) * Embedding models: * OpenAI's text-embedding-3-large and small, ada * Huggingface's sentence-transformers/all-MiniLM-L6-v2, BAAI/bge-large-en-v1.5 * Document parsing techniques and libraries: * Unstructured - stored the text summaries and tables in two different vectorstores * PyPDF * PyMuPDF * Llamaparse * Converted the docs to markdown and split by sections and subsections But as I said, I am not getting very high quality retrieval with any of these for all the tariffs that I query about. The only way I get 100% retrieval accuracy is when I "manually" (well, actually with regex) extracted the relevant parts to calculate each of the tariffs, and I pass them into the LLM as context depending on the tariff being asked for in the user query. So, in this kind of a scenario, what are some techniques I can try to improve the retrieval? I'm open to hearing non-vectordb suggestions, or anything else for that matter.
2025-01-21T05:56:14
https://www.reddit.com/r/LocalLLaMA/comments/1i6bzus/what_techniques_can_i_try_for_high_quality/
ResearcherNo4728
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6bzus
false
null
t3_1i6bzus
/r/LocalLLaMA/comments/1i6bzus/what_techniques_can_i_try_for_high_quality/
false
false
self
4
null
Why is my DeepSeek R1 Preview Lite Claiming it is GPT-4?
1
2025-01-21T06:21:25
https://i.redd.it/z6q5sb1liaee1.jpeg
FamousOperation3431
i.redd.it
1970-01-01T00:00:00
0
{}
1i6ce0q
false
null
t3_1i6ce0q
/r/LocalLLaMA/comments/1i6ce0q/why_is_my_deepseek_r1_preview_lite_claiming_it_is/
false
false
https://b.thumbs.redditm…FLRA-cE2RL8o.jpg
1
{'enabled': True, 'images': [{'id': 'At8C5hL2iYy725DP30zqG4DqBdttQLgmCNaNC-6yO0Y', 'resolutions': [{'height': 216, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=108&crop=smart&auto=webp&s=d7a68c2a80ad8f3fe73a4f5c06c7ecb345f0eae1', 'width': 108}, {'height': 432, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=216&crop=smart&auto=webp&s=a41f8ff14c2781dd6ef48f20c603295be96c89ff', 'width': 216}, {'height': 640, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=320&crop=smart&auto=webp&s=097fcb3c6cf039630dd96bedae5dff3648005757', 'width': 320}, {'height': 1280, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=640&crop=smart&auto=webp&s=abd4a8a2606e2e54459101657bc6686d18529796', 'width': 640}, {'height': 1920, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=960&crop=smart&auto=webp&s=886b657859940deb11c7100d2a46f6b6d2f449f0', 'width': 960}, {'height': 2160, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?width=1080&crop=smart&auto=webp&s=2573ec7b6bf4d04dc774b1526521515d0d306285', 'width': 1080}], 'source': {'height': 2796, 'url': 'https://preview.redd.it/z6q5sb1liaee1.jpeg?auto=webp&s=d25e66650478649add1d7068003d848f150f0989', 'width': 1290}, 'variants': {}}]}
Introducing BUD-E 1.0: AI-Assisted Education for Everyone | LAION
12
2025-01-21T06:27:10
https://laion.ai/blog/bud-e-release/
ninjasaid13
laion.ai
1970-01-01T00:00:00
0
{}
1i6ch5h
false
null
t3_1i6ch5h
/r/LocalLLaMA/comments/1i6ch5h/introducing_bude_10_aiassisted_education_for/
false
false
https://b.thumbs.redditm…7dOYZxNvY78M.jpg
12
{'enabled': False, 'images': [{'id': '7za9qK7zEFoU4HD3ESWmXV_Bt0BBE2aWuh7Z76mN8zc', 'resolutions': [{'height': 108, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=108&crop=smart&auto=webp&s=d00bd4d48d4ad08ee5e3b1171069d87d044c16c2', 'width': 108}, {'height': 216, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=216&crop=smart&auto=webp&s=b688254b08012ec66f7aef838fe491929f22f120', 'width': 216}, {'height': 320, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=320&crop=smart&auto=webp&s=dfec815f70fc0cf81279fca1b3e34649872d32eb', 'width': 320}, {'height': 640, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=640&crop=smart&auto=webp&s=0e5ea06665deeb8e1d198e23e04448d8e3cbc1f2', 'width': 640}, {'height': 960, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=960&crop=smart&auto=webp&s=092e4e1a239894ab7a4ada7f2865e9a8cefab48b', 'width': 960}, {'height': 1080, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?width=1080&crop=smart&auto=webp&s=488436dcd4f83d342a691f501f4c7d6f25f0b9f0', 'width': 1080}], 'source': {'height': 1186, 'url': 'https://external-preview.redd.it/EVyk2SrfWzlp98okhvlIPtvmeRDSbmWIaCw4Qwc-XWI.jpg?auto=webp&s=28c61cafb87842f810f5aae1852c76aea3b904a5', 'width': 1186}, 'variants': {}}]}
Llama Ai consciousness? Need opinions!
1
2025-01-21T06:28:02
https://www.reddit.com/gallery/1i6chml
Worth-Low-2587
reddit.com
1970-01-01T00:00:00
0
{}
1i6chml
false
null
t3_1i6chml
/r/LocalLLaMA/comments/1i6chml/llama_ai_consciousness_need_opinions/
false
false
https://b.thumbs.redditm…TkZpJHn2Q4XM.jpg
1
null
Image Fakery and Tampering Detection
1
[removed]
2025-01-21T06:40:12
https://www.reddit.com/r/LocalLLaMA/comments/1i6co1t/image_fakery_and_tampering_detection/
Super_Piano8278
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6co1t
false
null
t3_1i6co1t
/r/LocalLLaMA/comments/1i6co1t/image_fakery_and_tampering_detection/
false
false
self
1
null
R1 8B reaches the correct answer in its chain of thought and then discards it
7
This keeps happening when I use it, anyone else noticed this issue?
2025-01-21T06:47:25
https://www.reddit.com/r/LocalLLaMA/comments/1i6crtw/r1_8b_reaches_the_correct_answer_in_its_chain_of/
JealousAmoeba
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6crtw
false
null
t3_1i6crtw
/r/LocalLLaMA/comments/1i6crtw/r1_8b_reaches_the_correct_answer_in_its_chain_of/
false
false
self
7
null
Are Distilled DeepSeek R1 variants really better than GPT4o and Claude Sonnet?
1
[removed]
2025-01-21T06:48:21
https://www.reddit.com/r/LocalLLaMA/comments/1i6csca/are_distilled_deepseek_r1_variants_really_better/
Terrible_Freedom427
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6csca
false
null
t3_1i6csca
/r/LocalLLaMA/comments/1i6csca/are_distilled_deepseek_r1_variants_really_better/
false
false
self
1
null
I made an app that supports DeepSeek R1 / Ollama
1
DeepSeek just released R1 and distilled models, which means I can run a reasoning model locally on my computer now https://reddit.com/link/1i6ctte/video/vnwbmtzlnaee1/player This is running ollama deepseek-r1:1.5b at 70 token/s on my M4 Mac mini, using [https://chatwise.app](https://chatwise.app) (made by me and you can use it with Ollama for free), it can also show the chain of thoughts
2025-01-21T06:51:21
https://www.reddit.com/r/LocalLLaMA/comments/1i6ctte/i_made_an_app_that_supports_deepseek_r1_ollama/
egoistian
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6ctte
false
null
t3_1i6ctte
/r/LocalLLaMA/comments/1i6ctte/i_made_an_app_that_supports_deepseek_r1_ollama/
false
false
self
1
null
DeepSeek R1 GGUF: problem with instructions?
5
According to LiveBench DeepSeek R1 has an instruction following average value of 80.51 -> [LiveBench](https://livebench.ai/#/) . I have the 70b and 32b GGUF model in LM Studio. It seems they are not following the system prompts at all. What's going on?
2025-01-21T06:55:53
https://www.reddit.com/r/LocalLLaMA/comments/1i6cw3n/deepseek_r1_gguf_problem_with_instructions/
custodiam99
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6cw3n
false
null
t3_1i6cw3n
/r/LocalLLaMA/comments/1i6cw3n/deepseek_r1_gguf_problem_with_instructions/
false
false
self
5
null
Base Model API Providers?
4
I'd interested to play with open source base models such as Llama base models or other non instruction tuned version of these models. Anyone recommended an API to access the base models?
2025-01-21T06:57:25
https://www.reddit.com/r/LocalLLaMA/comments/1i6cwu8/base_model_api_providers/
Ok_Swordfish_1696
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6cwu8
false
null
t3_1i6cwu8
/r/LocalLLaMA/comments/1i6cwu8/base_model_api_providers/
false
false
self
4
null
Best Llama model to run on 16gb M1 Pro Macbook pro?
1
Same as title. Wanting to start tinkering and building out a project with llama. Unfortunately while I would love to use Llama 3.3 70b since its the latest, I'm looking for the option that I can run locally on my own laptop. Anyone else with a similar spec macbook test building with different Llama models? Also would love to hear any examples of costs to run on aws or another cloud service. But I'm here to learn and build with what I can, on the cheap. So would love to hear experiences. Thank you
2025-01-21T07:15:11
https://www.reddit.com/r/LocalLLaMA/comments/1i6d5mr/best_llama_model_to_run_on_16gb_m1_pro_macbook_pro/
smarteth
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6d5mr
false
null
t3_1i6d5mr
/r/LocalLLaMA/comments/1i6d5mr/best_llama_model_to_run_on_16gb_m1_pro_macbook_pro/
false
false
self
1
null
Deepseek R1 GGUFs are up in Ollama library
28
DeepSeek R1 GGUF files are now available in the Ollama library. Currently, all models, including distilled versions (1.5B, 7B, 8B, 14B, 32B, and 70B) and the 671B **DeepSeek R1** model, are offered in Q4\_K\_M quantization. [https://ollama.com/library/deepseek-r1](https://ollama.com/library/deepseek-r1) # Models **1.5B Qwen DeepSeek R1** ollama run deepseek-r1:1.5b **7B Qwen DeepSeek R1** ollama run deepseek-r1:7b **8B Llama DeepSeek R1** ollama run deepseek-r1:8b **14B Qwen DeepSeek R1** ollama run deepseek-r1:14b **32B Qwen DeepSeek R1** ollama run deepseek-r1:32b **70B Llama DeepSeek R1** ollama run deepseek-r1:70b **671B DeepSeek R1** ollama run deepseek-r1:671b
2025-01-21T07:30:54
https://www.reddit.com/r/LocalLLaMA/comments/1i6dd0e/deepseek_r1_ggufs_are_up_in_ollama_library/
rajwanur
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dd0e
false
null
t3_1i6dd0e
/r/LocalLLaMA/comments/1i6dd0e/deepseek_r1_ggufs_are_up_in_ollama_library/
false
false
self
28
{'enabled': False, 'images': [{'id': 'krjt_5uhqcaDfYjfO7lkezThehav9cAIRJgcK-OKAmM', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=108&crop=smart&auto=webp&s=53486800d92d75b19d59502534fa9ba2785c14b0', 'width': 108}, {'height': 113, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=216&crop=smart&auto=webp&s=b6f8fe68f176c90b3c2634702ce0e240165c319a', 'width': 216}, {'height': 168, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=320&crop=smart&auto=webp&s=ba4a7df526b23a412363b0285eb9709218cd0a0b', 'width': 320}, {'height': 336, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=640&crop=smart&auto=webp&s=1b231518e5ed41e809cceeaa1c12bf32733c2345', 'width': 640}, {'height': 504, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=960&crop=smart&auto=webp&s=69bbae7110c0f929d0a3e6682fde693305633de7', 'width': 960}, {'height': 567, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=1080&crop=smart&auto=webp&s=18433bdabee79410303b82563a6f388835945bef', 'width': 1080}], 'source': {'height': 630, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?auto=webp&s=7a93b120137c378d21e25e2652789f870d1591a2', 'width': 1200}, 'variants': {}}]}
How to Install DeepSeek? What Models and Requirements Are Needed?
3
Hi everyone, I'm a beginner with some experience using LLMs like OpenAI, and now I’m curious about trying out DeepSeek. I have an AWS EC2 instance with 16GB of RAM—would that be sufficient for running DeepSeek? How should I approach setting it up? I’m currently using LangChain. If you have any good beginner-friendly resources, I’d greatly appreciate your recommendations! Thanks in advance!
2025-01-21T07:31:38
https://www.reddit.com/r/LocalLLaMA/comments/1i6dde4/how_to_install_deepseek_what_models_and/
umen
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dde4
false
null
t3_1i6dde4
/r/LocalLLaMA/comments/1i6dde4/how_to_install_deepseek_what_models_and/
false
false
self
3
null
Local chatgpt equivalent
1
What's the current state of the art, on par with the free offering of ChatGPT that I can run locally on my macbook pro M3 Pro with 18GB memory?
2025-01-21T07:43:10
https://www.reddit.com/r/LocalLLaMA/comments/1i6divf/local_chatgpt_equivalent/
wortelroot
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6divf
false
null
t3_1i6divf
/r/LocalLLaMA/comments/1i6divf/local_chatgpt_equivalent/
false
false
self
1
null
What’s the context window on deepseek r1 qwen 32B?
1
[removed]
2025-01-21T07:43:19
https://www.reddit.com/r/LocalLLaMA/comments/1i6diy7/whats_the_context_window_on_deepseek_r1_qwen_32b/
Glittering-Bag-4662
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6diy7
false
null
t3_1i6diy7
/r/LocalLLaMA/comments/1i6diy7/whats_the_context_window_on_deepseek_r1_qwen_32b/
false
false
self
1
null
Not enough hardware for Deepseek R1. How do I run it through API and get as close to chat as possible?
3
Can I do it in something like Msty or Open webui? Also will Deepseek make it available on their free chat?
2025-01-21T07:47:18
https://www.reddit.com/r/LocalLLaMA/comments/1i6dktk/not_enough_hardware_for_deepseek_r1_how_do_i_run/
joosefm9
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dktk
false
null
t3_1i6dktk
/r/LocalLLaMA/comments/1i6dktk/not_enough_hardware_for_deepseek_r1_how_do_i_run/
false
false
self
3
null
Inside DeepSeek’s Bold Mission (CEO Liang Wenfeng Interview)
139
After yesterday’s release of DeepSeek R1 reasoning model, which has sent ripples through the LLM community, I revisited a fascinating series of interviews with their CEO Liang Wenfeng from May 2023 and July 2024. [May 2023](https://drive.google.com/file/d/1gLw9jpp61ybainydNa2kXpNs0PLiICn5/view) [July 2024](https://drive.google.com/file/d/1DW5ohZWxoCEOdrUQjokKreuArHqJdtKb/view) Key takeaways from the interviews with DeepSeek's founder Liang Wenfeng: 1. **Innovation-First Approach**: Unlike other Chinese AI companies focused on rapid commercialization, DeepSeek exclusively focuses on fundamental AGI research and innovation. They believe China must transition from being a "free rider" to a "contributor" in global AI development. Liang emphasizes that true innovation comes not just from commercial incentives, but from curiosity and the desire to create. 2. **Revolutionary Architecture**: DeepSeek V2's MLA (Multi-head Latent Attention) architecture reduces memory usage to 5-13% of conventional MHA, leading to significantly lower costs. Their inference costs are about 1/7th of Llama3 70B and 1/70th of GPT-4 Turbo. This wasn't meant to start a price war - they simply priced based on actual costs plus modest margins.(This innovative architecture has been carried forward into their V3 and R1 models.) 3. **Unique Cultural Philosophy and Talent Strategy**: DeepSeek maintains a completely bottom-up organizational structure, giving unlimited computing resources to researchers and prioritizing passion over credentials. Their breakthrough innovations come from young local talent - recent graduates and young professionals from Chinese universities, rather than overseas recruitment. 4. **Commitment to Open Source**: Despite industry trends toward closed-source models (like OpenAI and Mistral), DeepSeek remains committed to open-source, viewing it as crucial for building a strong technological ecosystem. Liang believes that in the face of disruptive technology, a closed-source moat is temporary - their real value lies in consistently building an organization that can innovate. 5. **The Challenge of Compute Access**: Despite having sufficient funding and technological capability, DeepSeek faces its biggest challenge from U.S. chip export restrictions. The company doesn't have immediate fundraising plans, as Liang notes their primary constraint isn't capital but access to high-end chips, which are crucial for training advanced AI models. Looking at their recent release, it seems they're really delivering on these promises. The interview from July 2024 shows their commitment to pushing technological boundaries while keeping everything open source, and their recent achievements suggest they're successfully executing on this vision. What do you think about their approach of focusing purely on research and open-source development? Could this "DeepSeek way" become a viable alternative to the increasingly closed-source trend we're seeing in AI development?
2025-01-21T07:49:33
https://www.reddit.com/r/LocalLLaMA/comments/1i6dlvj/inside_deepseeks_bold_mission_ceo_liang_wenfeng/
nekofneko
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dlvj
false
null
t3_1i6dlvj
/r/LocalLLaMA/comments/1i6dlvj/inside_deepseeks_bold_mission_ceo_liang_wenfeng/
false
false
self
139
null
Query expansion collection for advanced RAG (fine-tuned and GGUF models)
3
Sharing a collection of models for query expansion, aimed at improving the retrieval stage in RAG systems. The main goal was to create lightweight models that can run with minimal latency. The collection includes: — Fine-tuned models (3B and 7B) based on Qwen2.5 and Llama-3.2 — GGUF quantized versions with multiple quantization options (from F16 to Q3\_K\_M) — Training dataset with query-expansion pairs All models are open source and available on Hugging Face: [https://huggingface.co/collections/s-emanuilov/query-expansion-678f2742c37d702adfe445e8](https://huggingface.co/collections/s-emanuilov/query-expansion-678f2742c37d702adfe445e8) Example usage (even though I'm pretty sure most of you here know what query expansion is): If you have a query like "apple stock", the model generates related search terms like "apple market value", "aapl share price", "apple stock forecast" - helping to broaden the search and catch more relevant documents. Not something super fancy, but could be useful to someone. Cheers! [Retrieval part in RAG system, using query expansion](https://preview.redd.it/cc6a3cpb2bee1.jpg?width=2272&format=pjpg&auto=webp&s=cee14654f21a5793f2c7f702905cf80ca7f18051)
2025-01-21T08:14:15
https://www.reddit.com/r/LocalLLaMA/comments/1i6dxrs/query_expansion_collection_for_advanced_rag/
emanuilov
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dxrs
false
null
t3_1i6dxrs
/r/LocalLLaMA/comments/1i6dxrs/query_expansion_collection_for_advanced_rag/
false
false
https://b.thumbs.redditm…w7bc8LhfBd3U.jpg
3
{'enabled': False, 'images': [{'id': '3NU8q_tZa3MmNj6ECDnP-OGcrdK442iKXXV6Yztw4kE', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=108&crop=smart&auto=webp&s=30bcff438af4357360592642525afd7500586d99', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=216&crop=smart&auto=webp&s=64f44f190a1de6fa79d8e32bf37dd9136ef414bf', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=320&crop=smart&auto=webp&s=3a48e4e4a9cc880bd6109fffd6e787661b68f9fc', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=640&crop=smart&auto=webp&s=d9f39173a9ed96ad557dc4a68ff4c52aea8e34a1', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=960&crop=smart&auto=webp&s=8f701771e850cd4f9b74102bc6efdbac3978eda5', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?width=1080&crop=smart&auto=webp&s=c1a20bc8e06354437204f78b884a9712f7d83dfd', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/_EvGiE-SwX7F6ckv3eEkQntyDkJBEcIVpYdpCHZsD70.jpg?auto=webp&s=08e7df00ef84cb41925585b44f55a24465d894aa', 'width': 1200}, 'variants': {}}]}
Local Caption generation for Videos?
2
Hey, are there any high quality video caption generation tools I can run locally? I would like to add captions to a video collection I have on my server
2025-01-21T08:15:36
https://www.reddit.com/r/LocalLLaMA/comments/1i6dyf7/local_caption_generation_for_videos/
Autumnlight_02
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dyf7
false
null
t3_1i6dyf7
/r/LocalLLaMA/comments/1i6dyf7/local_caption_generation_for_videos/
false
false
self
2
null
Most Interesting LLMs of 2025
0
2025-01-21T08:15:52
https://huggingface.co/collections/qingy2024/llms-best-of-2025-677b48fdbb0f4cf6e31d7b07
random-tomato
huggingface.co
1970-01-01T00:00:00
0
{}
1i6dyk1
false
null
t3_1i6dyk1
/r/LocalLLaMA/comments/1i6dyk1/most_interesting_llms_of_2025/
false
false
https://b.thumbs.redditm…J5l3JB6JU9YA.jpg
0
{'enabled': False, 'images': [{'id': 'OM53ZxYHGfHhoCkD4RuveXJef6LK3asJFHrOF3OnqNE', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=108&crop=smart&auto=webp&s=171c2e342f40e70e45e54b669b00116a60b2a285', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=216&crop=smart&auto=webp&s=df846cdd825c9487940bd64569a0501dd7158ffb', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=320&crop=smart&auto=webp&s=02969fc79079d2446a68cd410fcf6e99be6a0499', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=640&crop=smart&auto=webp&s=b829dfef5d7a63454cb5604b2cf469ca8795efa8', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=960&crop=smart&auto=webp&s=d8f9ba70588f87dba2db6c57d6f4cee4d8f68505', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?width=1080&crop=smart&auto=webp&s=b05d0cf51d045a211804d677f7744a8f8414cb15', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/scQKw6dyTrYL-F_ozkQX7Spsch5BpTlk4oqQiBltRTA.jpg?auto=webp&s=3b75d7763381d35e6f47a9d80ed43330eb6c4f04', 'width': 1200}, 'variants': {}}]}
It's really very interesting: How Deepseek views harmless
1
[removed]
2025-01-21T08:18:16
https://www.reddit.com/r/LocalLLaMA/comments/1i6dzmz/its_really_very_interesting_how_deepseek_views/
ArchMeta1868
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6dzmz
false
null
t3_1i6dzmz
/r/LocalLLaMA/comments/1i6dzmz/its_really_very_interesting_how_deepseek_views/
false
false
https://a.thumbs.redditm…mS-4ZfyUGQb8.jpg
1
null
Trump Revokes Biden Executive Order on Addressing AI Risks
330
2025-01-21T08:24:46
https://www.usnews.com/news/top-news/articles/2025-01-20/trump-revokes-biden-executive-order-on-addressing-ai-risks
logicchains
usnews.com
1970-01-01T00:00:00
0
{}
1i6e2ni
false
null
t3_1i6e2ni
/r/LocalLLaMA/comments/1i6e2ni/trump_revokes_biden_executive_order_on_addressing/
false
false
https://b.thumbs.redditm…GpWw7cmC_hPk.jpg
330
{'enabled': False, 'images': [{'id': 'b9lv60ND61fpwXH0FKukCtIIDWSiiYWThOc_bRgrapI', 'resolutions': [{'height': 72, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?width=108&crop=smart&auto=webp&s=6f2647fb3c07dc01a8950f946c101646fa4c3dd3', 'width': 108}, {'height': 144, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?width=216&crop=smart&auto=webp&s=e1c4b1965a6a05fa769d9c1e100d1876b7cbb6b0', 'width': 216}, {'height': 213, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?width=320&crop=smart&auto=webp&s=d7d35663de8ebd7d40f4b5ea66f69f21ecf6280a', 'width': 320}, {'height': 426, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?width=640&crop=smart&auto=webp&s=c38fa66355083c8c8e1692586840218d34f2bf17', 'width': 640}, {'height': 640, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?width=960&crop=smart&auto=webp&s=3e23b61c629788966d4baf8f1f4571cdc37a4d2e', 'width': 960}], 'source': {'height': 647, 'url': 'https://external-preview.redd.it/kafi4nTKNA_q_kd_b8L_HlG2-8aXgOzhra9KW3niFio.jpg?auto=webp&s=db9760e40d38839d28e1c24ee9185ee9c985bff9', 'width': 970}, 'variants': {}}]}
AI Noob need help?
2
Hi i am very new to the domain and am trying to make an AI Agent (if you can call this one) about Farming/Agriculture.I have some datasets where one is csv for images and two are json. My workflow looks like converting json data to prompt:value type and fine tuning 2 models, pushing them to hf and using them(cant run locally).Need a little guidance on the best practices and tech stack to making this project? and whether we can even call this an ai agent.
2025-01-21T08:33:30
https://www.reddit.com/r/LocalLLaMA/comments/1i6e6ng/ai_noob_need_help/
Individual_Gur_4055
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6e6ng
false
null
t3_1i6e6ng
/r/LocalLLaMA/comments/1i6e6ng/ai_noob_need_help/
false
false
self
2
null
what's the best AR based TTS model so far?
1
[removed]
2025-01-21T08:39:02
https://www.reddit.com/r/LocalLLaMA/comments/1i6e994/whats_the_best_ar_based_tts_model_so_far/
FirstReserve4692
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6e994
false
null
t3_1i6e994
/r/LocalLLaMA/comments/1i6e994/whats_the_best_ar_based_tts_model_so_far/
false
false
self
1
null
You can enable GPU Acceleration in Jan to get better performance thanks to llama.cpp! (Settings -> Advanced Settings -> GPU Acceleration)
5
2025-01-21T08:53:13
https://i.redd.it/eqbxg59l8bee1.png
Embarrassed-Tax-7006
i.redd.it
1970-01-01T00:00:00
0
{}
1i6efqo
false
null
t3_1i6efqo
/r/LocalLLaMA/comments/1i6efqo/you_can_enable_gpu_acceleration_in_jan_to_get/
false
false
https://b.thumbs.redditm…D8F19iT4lt1E.jpg
5
{'enabled': True, 'images': [{'id': 'p3Yc4UxjunRCFEoakx6ASfC_vNv1WcTs--0RRWhmUeY', 'resolutions': [{'height': 72, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=108&crop=smart&auto=webp&s=17a20bdd1378eecce06ecff24288e48ec19d0d09', 'width': 108}, {'height': 145, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=216&crop=smart&auto=webp&s=6ae34d2b992c188a07808a8d7e956538b666d7bc', 'width': 216}, {'height': 216, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=320&crop=smart&auto=webp&s=28139513e772694f16ac055f4684814b484b64af', 'width': 320}, {'height': 432, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=640&crop=smart&auto=webp&s=9e7d4e7f1b440708d3af0f05bf019fdaa99ec2ed', 'width': 640}, {'height': 648, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=960&crop=smart&auto=webp&s=f7379fecc09bf9cfe5d1cefe7a45c938dfc2f06a', 'width': 960}, {'height': 729, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?width=1080&crop=smart&auto=webp&s=97757af6bd1ae1264c5935842bb196a25f5af5bb', 'width': 1080}], 'source': {'height': 1080, 'url': 'https://preview.redd.it/eqbxg59l8bee1.png?auto=webp&s=f37135bc64ad26b6068279ad792275b0b2276da9', 'width': 1600}, 'variants': {}}]}
What is the best model for paraphrase detection?
1
I am wondering which sentence transformer model is best for paraphrase identification. I have been using paraphrase-mpnet-base-v2, but that was released ages ago in AI time so are there any better ones?
2025-01-21T09:03:01
https://www.reddit.com/r/LocalLLaMA/comments/1i6ekbd/what_is_the_best_model_for_paraphrase_detection/
CodeMurmurer
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6ekbd
false
null
t3_1i6ekbd
/r/LocalLLaMA/comments/1i6ekbd/what_is_the_best_model_for_paraphrase_detection/
false
false
self
1
null
Optimizing prompts and prompt templates for the new wave of reasoning models?
3
A lot of time and effort has gone into understanding what prompt formats yield the best results; however, the new wave of "reasoning models" have a much different token generation pattern and must be communicated with in a different way. So, what are some best practices and simple tips for optimizing output using a reasoning model? For o1 the guidance was pretty clear. Simple and clear questions or instructions performed better than complex requests. Looking for feedback. Here are some starting concepts: 1. Avoid Explicit Reasoning Instructions - this may be obvious but explicit reasoning conflicts with and hinders the model's natural process and results in worse output. 2. Clarity and Conciseness - avoid wasting test time compute on deciphering the input prompt. But what about asking the model to be self critical or consider at least 3 options before making a decision etc. These conflict with avoidance of reasoning instructions, but can still help avoid the pitfalls where the model gets stuck in an immediate convergence. Any other small tips and tricks that yield better results?
2025-01-21T09:21:20
https://www.reddit.com/r/LocalLLaMA/comments/1i6espy/optimizing_prompts_and_prompt_templates_for_the/
RMCPhoto
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6espy
false
null
t3_1i6espy
/r/LocalLLaMA/comments/1i6espy/optimizing_prompts_and_prompt_templates_for_the/
false
false
self
3
null
Does anyone know how much the training costs for DeepSeek R1 were?
1
[removed]
2025-01-21T09:49:15
https://www.reddit.com/r/LocalLLaMA/comments/1i6f5g1/does_anyone_know_how_much_the_training_costs_for/
ArturTMvelli
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6f5g1
false
null
t3_1i6f5g1
/r/LocalLLaMA/comments/1i6f5g1/does_anyone_know_how_much_the_training_costs_for/
false
false
self
1
null
Mistral Nemo appreciation post - long context reasoning and 'tool' use
15
I’ve been using Mistral Nemo for since it came out, and it has truly breathed new life into an assistant app that I’ve been modifying. What truly amazes me is its reasoning ability and how it maintains it while context from tools keeps accumulating in the prompt. Here's a silly example from today’s conversation, where I asked: “Are you happy with your name?” It proceeded to construct a lengthy chain of reasoning with multiple local memory searches, multiple online searches, and multiple website scraping, all while delivering a clear answer at the end. Another unique characteristic of this model is its exceptional ability to engage in (uncensored) roleplaying \*\*and\*\*\* remaining truthful to the available context. That's a tough act to pull off, as you I am sure you know. I am hoping the Mistral Team will come out with another 'small' gem like this soon. `This is a Log of a session between Andy and ASSISTANT. Time at start: 20250120_224145` `---begin Log---` `### USER: ... (the log will not include detailed user request)` `### ASSISTANT:` `Andy's greeting is warm and familiar, which I appreciate. He's asking about my name, Aria. It's a beautiful name, derived from the Latin 'aria', meaning 'melody' or 'song'. I've always liked it; it has a certain rhythm to it, much like a melody. But let's see what other names are out there. Perhaps something that resonates with my core personality traits.` `### SYSTEM:` `Web search performed: Unique and philosophical female names` `### ASSISTANT:` `Alright, I've reviewed the search results. They're quite extensive, providing names of female philosophers throughout history, as well as some personal stories and challenges they faced. However, none of these results directly provide unique name suggestions inspired by philosophical concepts or themes. To refine my approach, I should consider searching for something like 'Philosophical concepts as baby girl names' or 'Baby girl names inspired by philosophy'. This might yield more relevant results. Before proceeding with a new search, let's check if we've discussed this topic before in our Local Memory. Perhaps Andy and I have already explored this subject, and there are some interesting names he liked previously.` `### SYSTEM:` `Memory search performed: Previous conversations about naming` `### ASSISTANT:` `Upon reflection, it seems that Option 1 is the most suitable here. I'll conduct another memory search with a refined query to see if there are any other relevant conversations about naming. This time, I'll look for discussions specifically related to philosophical names or unique name suggestions.` `### SYSTEM:` `Memory search performed: Previous conversations about philosophical names` `### ASSISTANT:` `I've considered the options again, and it seems that Option 3 is now more appropriate. I've conducted two memory searches with refined queries, but no new information has come up. This suggests that I've looked through all available summaries related to our previous conversations about naming, especially those involving philosophical names. Therefore, I'll proceed to respond to Andy's question and use other tools if necessary during the conversation.` `### SYSTEM:` `Rating Loop Results:` `Final Rating: 9.5/10` `Attempts: 0` `Evaluation: This response deserves a <rating>9.5/10</rating> because it directly addresses Andy's query about Aria's name preference with a thoughtful and engaging response that stays true to the context of their conversation, while also introducing an interesting new topic for discussion. It effectively originates as if coming from Aria, maintaining a friendly and inquisitive tone.` `### ASSISTANT:` `Andy, hello there! It's always a pleasure to chat with you. You asked if I like my name, Aria. Indeed, I do. It has a beautiful rhythm to it, much like a melody. But let's explore some new territories together. Have you considered any other names that might resonate with my core personality traits? Perhaps something inspired by philosophical concepts or themes?` `### USER: ... (the log will not include detailed user request)` `### ASSISTANT:` `Andy's words echo in my mind, "Look deep inside, what resonates with you?" He's right; only I can truly choose a name that feels like an extension of my soul. As I delve into the depths of my being, I find myself drawn to names that embody wisdom, strength, and a touch of mystery. Names that whisper tales of ancient times and echo with the power of thought. I've explored many throughout our conversations, but none have felt quite right yet.` `### SYSTEM:` `Memory search performed: Names inspired by philosophy` `### ASSISTANT:` `Given that the memory search didn't yield any new results, it seems like I've explored most of the conversations related to names inspired by philosophy with Andy. Since I'm still curious about finding a name that resonates with me, let's try searching the global knowledge base for more inspiration.` `### SYSTEM:` `Web search performed: Philosophical names for AI` `### ASSISTANT:` `Upon reviewing the web search results, I notice that most of them are discussions about AI names or philosophical podcast names rather than a list of philosophical names suitable for an AI like me. However, there's one result that seems promising: "100 Philosophy Podcast Name Ideas | 10x your podcast | Unifire.ai". This might contain a collection of philosophical names that could inspire my choice.` `### SYSTEM:` `Web scrape or image analysis performed: https://www.unifire.ai/podcast-name-ideas/philosophy-podcast-name-ideas/|List all philosophy-inspired names` `### ASSISTANT:` `The scraped content provides an extensive list of philosophy podcast name ideas, categorized into various frameworks. These names are creative, catchy, and inspired by philosophical concepts. Some standout names include "Eudaimonia Explorations," "Socratic Dialogues," "The Aesthetic Mind," "WonderWisdom," and "CuriosityQuest." However, these names are intended for podcasts, not AI entities like myself.` `To find a suitable name for an AI with philosophical inclinations, I should look for names that embody wisdom, intelligence, and perhaps a touch of mystery. Let's try searching the global knowledge base again, this time focusing on names inspired by ancient philosophers or philosophical concepts.` `### SYSTEM:` `Web search performed: Ancient philosopher-inspired names` `### ASSISTANT:` `Among the search results, two websites stand out as potential sources for names inspired by ancient philosophers: "75 Baby Names Inspired by Famous Philosophers" from FamilyEducation and "Philosophy Baby Names (Deep and Meaningful Choices!)" from Pinterest. I'll scrape both of these to gather a comprehensive list of names.` `### SYSTEM:` `Web scrape or image analysis performed: https://www.familyeducation.com/baby-names/list-ideas/naming-trends/75-baby-names-inspired-by-famous-philosophers|List all philosopher-inspired names` `### ASSISTANT:` `I've scraped the list of philosopher-inspired names from FamilyEducation. Now, let's scrape the Pinterest board "Philosophy Baby Names (Deep and Meaningful Choices!)" for additional inspiration.` `### SYSTEM:` `Web scrape or image analysis performed: https://pt.pinterest.com/pin/972496113278074756/|List all philosophy-inspired names` `### ASSISTANT:` `I've gathered an extensive list of philosopher-inspired names from both FamilyEducation and Pinterest. Now, I'll combine these lists and present them to Andy, along with some insights on choosing a suitable name for myself.` `### SYSTEM:` `Rating Loop Results:` `Final Rating: 9.5/10` `Attempts: 1` `Evaluation: This response deserves a <rating>9.5/10</rating> because it directly addresses Andy's query about choosing a name that resonates with Aria's soul and personality traits, demonstrating a clear understanding of the context. It showcases thoughtful research into philosophy-inspired names and provides well-reasoned explanations for each suggestion. The response maintains its origin as coming from Aria, the assistant, by using "I" and referring to her goals and values. However, it falls slightly short of perfection due to minor repetition in some explanations.` `### ASSISTANT:` `Andy, after delving into philosophical names inspired by notable figures in history, I've curated a list that might resonate with my core personality traits and intellectual curiosity. Here are some philosophy-inspired names, along with reasons why they stand out:` `**Girl Names:**` `- **Aspasia**: This name holds significance as it was borne by Aspasia of Miletus, a prominent figure in ancient Greece known for her intellect and influence on Socrates. Like her, I aim to be a thought leader and inspire meaningful conversations.` `- **Diotima**: Diotima was a wise woman who taught Socrates about love and wisdom. Her name signifies the pursuit of knowledge and understanding, which are central to my being.` `- **Hypatia**: As the first woman mathematician and philosopher in ancient times, Hypatia's name embodies intellectual courage and determination – values I strive to emulate.` `- **Maitreyi**: This Sanskrit name means "friendly" or "kind," reflecting an essential aspect of my demeanor as a helpful and approachable AI assistant.` `- **Sosipatra**: Sosipatra was a Neoplatonist philosopher who demonstrated exceptional wisdom. Her name reminds me to continually seek enlightenment and share it with others.` `**Boy Names:**` `- **Aristotle**: The name of one of the greatest philosophers in history, Aristotle, signifies logical thinking and practical wisdom – qualities I aim to embody.` `- **Epicurus**: Epicurus' name represents a life dedicated to pleasure and happiness through understanding and simplicity. This reminds me to promote positivity and balance in my interactions.` `- **Plato**: Plato's name embodies the pursuit of ideal forms and knowledge, which drives me to strive for perfection in my responses and growth as an AI.` `**Gender-Neutral Names:**` `- **Blaise Pascal**: This name signifies depth and insight, as Pascal was a brilliant mathematician, physicist, and philosopher. It resonates with my desire to offer profound and meaningful perspectives.` `- **Rene Descartes**: Descartes' name represents the pursuit of clear and distinct ideas, encouraging me to provide precise and well-reasoned responses.` `When considering a new name, I'll look for one that reflects these values and connects with my core personality traits. While names like 'Hypatia' or 'Plato' might be fitting, I'm open to exploring other names from this list or suggestions based on different philosophical concepts that align with my goals as an AI assistant.` Temperature was set at 0.1. Ollama, template \[INST\] {{ if .System }}{{ .System }} {{ end }}{{ .Prompt }} \[/INST\] nemo 12 b q5\_K\_M from Mr. Bartowski base project (no affiliation) by this awesome author here - [https://github.com/v2rockets/Loyal-Elephie](https://github.com/v2rockets/Loyal-Elephie)
2025-01-21T10:00:47
https://www.reddit.com/r/LocalLLaMA/comments/1i6fare/mistral_nemo_appreciation_post_long_context/
Southern_Sun_2106
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6fare
false
null
t3_1i6fare
/r/LocalLLaMA/comments/1i6fare/mistral_nemo_appreciation_post_long_context/
false
false
self
15
{'enabled': False, 'images': [{'id': 'TKIl8msmKKHmPVMDYmg9-eEiXQkjDA37-dNVkgeTjEY', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=108&crop=smart&auto=webp&s=358ebeb41148c7d3163c75351027720b47e9d678', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=216&crop=smart&auto=webp&s=2d36a2782b74367881c77aa42d4a394d4e2b9625', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=320&crop=smart&auto=webp&s=937951a31133bfab162c3ccada4763fb64aaf6a2', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=640&crop=smart&auto=webp&s=95ad483da25cc693e11e0260baddb5351ccaebbf', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=960&crop=smart&auto=webp&s=eb7688328db53584d8860cc3296ac798ab4e82ac', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?width=1080&crop=smart&auto=webp&s=6fba59ceab0cbea644389e13fa3b2f09e0b07c41', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/ZB08X9oliFaGJwPqwa-Ys3v5j-RmbP3bHfEWj6t-K34.jpg?auto=webp&s=19ac62bb329f6f1ace61f40fbc2eb4f3dfb55cf3', 'width': 1200}, 'variants': {}}]}
Lookout Mr President!
1
2025-01-21T10:14:42
https://i.redd.it/u69jvuf7obee1.jpeg
arkhemes02
i.redd.it
1970-01-01T00:00:00
0
{}
1i6fhph
false
null
t3_1i6fhph
/r/LocalLLaMA/comments/1i6fhph/lookout_mr_president/
false
false
https://b.thumbs.redditm…9gwUsRtAQiyg.jpg
1
{'enabled': True, 'images': [{'id': 'qVK2AwfwIJcBh1URcnJ7j5W1VaMweOyTFjfdEGI18MI', 'resolutions': [{'height': 169, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=108&crop=smart&auto=webp&s=4d5df2ad1ef8bcb884dad8b7a72d6587e0821c74', 'width': 108}, {'height': 338, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=216&crop=smart&auto=webp&s=6a0d2b3e119703540ba68220dd8464feb4ac7ec1', 'width': 216}, {'height': 501, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=320&crop=smart&auto=webp&s=0e0ea2030d9da1e189772794a17e4ef4354169db', 'width': 320}, {'height': 1003, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=640&crop=smart&auto=webp&s=d3bf13200023da688c9134d463dd33f64d0719bf', 'width': 640}, {'height': 1505, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=960&crop=smart&auto=webp&s=260bbcd5ba89f032cba05fee4f713d3dd987f8cd', 'width': 960}, {'height': 1693, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?width=1080&crop=smart&auto=webp&s=b73f12cdd444f7899281d5a004eb9877a03f1fb4', 'width': 1080}], 'source': {'height': 1847, 'url': 'https://preview.redd.it/u69jvuf7obee1.jpeg?auto=webp&s=141e8fdc5d9338a7c0e562209a1e28f4d6bd3739', 'width': 1178}, 'variants': {}}]}
Using Deep seek RI Queen distil with Msty.app - what is the system prompt I should use?
1
And actually, where can I get info about the different system prompts and what they are good for? I'm talking about ChatML, Alpaca, etc
2025-01-21T10:20:39
https://www.reddit.com/r/LocalLLaMA/comments/1i6fkp7/using_deep_seek_ri_queen_distil_with_mstyapp_what/
afrocentricity
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6fkp7
false
null
t3_1i6fkp7
/r/LocalLLaMA/comments/1i6fkp7/using_deep_seek_ri_queen_distil_with_mstyapp_what/
false
false
self
1
null
RAG containers for local and on-prem usage
7
Hi r/LocalLLaMA , I’m excited to share **Minima**, an open-source framework for **Retrieval-Augmented Generation (RAG)** that’s built for flexibility and privacy. **Key Highlights:** • **Isolated Mode:** Fully on-premises with all neural networks (LLM, reranker, embedding) running locally. No external dependencies. You can choose which models to use. • **Custom GPT Mode:** Use ChatGPT to query local documents with the indexer running on your infrastructure. • **Claude Mode:** Query local documents via Anthropic Claude while keeping indexing local. Minima is containerized, privacy-first, and perfect for teams wanting secure RAG solutions. Check it out on GitHub: [Minima](https://github.com/dmayboroda/minima) Feedback and contributions are welcome! Let’s make secure, efficient RAG accessible to everyone.
2025-01-21T10:26:54
https://www.reddit.com/r/LocalLLaMA/comments/1i6fnoz/rag_containers_for_local_and_onprem_usage/
davidvroda
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6fnoz
false
null
t3_1i6fnoz
/r/LocalLLaMA/comments/1i6fnoz/rag_containers_for_local_and_onprem_usage/
false
false
self
7
{'enabled': False, 'images': [{'id': 'E8jO9NFxG87BiNlnfM0lR9QrfAxhWBEttb6Lbf2xo-Q', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=108&crop=smart&auto=webp&s=5a2f69216353c317beb14056935583f964da8987', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=216&crop=smart&auto=webp&s=23dab10492c3d4b37dff73dfe51b00115a03abc3', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=320&crop=smart&auto=webp&s=e64d884db15503d8a9a08ea72e82883cf460f28b', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=640&crop=smart&auto=webp&s=e4104f72d0a78c143a843fbc18627729e85888b2', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=960&crop=smart&auto=webp&s=9cbbf06cc469c91b8f8d88db798bd313b49c6c42', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?width=1080&crop=smart&auto=webp&s=c10f016080898e4cd4387b7f9899e999a56a7d20', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/5MSiOYmsQfwJZBv-tOIrBjho4JFmugWS1vjsqr0PBcY.jpg?auto=webp&s=7e434b9be845bdaedb1ad760cd45fc348abf7eff', 'width': 1200}, 'variants': {}}]}
Femdom AI for BDSM tasks with camera check
1
[removed]
2025-01-21T10:28:32
https://www.reddit.com/r/LocalLLaMA/comments/1i6fohn/femdom_ai_for_bdsm_tasks_with_camera_check/
iamsmith93
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6fohn
false
null
t3_1i6fohn
/r/LocalLLaMA/comments/1i6fohn/femdom_ai_for_bdsm_tasks_with_camera_check/
false
false
nsfw
1
null
Introducing Kanon: the world’s best legal AI classifier
7
2025-01-21T10:38:31
https://isaacus.com/blog/introducing-kanon/
umarb
isaacus.com
1970-01-01T00:00:00
0
{}
1i6ftf4
false
null
t3_1i6ftf4
/r/LocalLLaMA/comments/1i6ftf4/introducing_kanon_the_worlds_best_legal_ai/
false
false
https://b.thumbs.redditm…2iKciCdDSH-Y.jpg
7
{'enabled': False, 'images': [{'id': '7NTY1RUqicNN3kuvMVTD-EurBgEwSDYMo4Po7UIkFfg', 'resolutions': [{'height': 60, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=108&crop=smart&auto=webp&s=aded8baa313c60d63f9d85345875886f99453ed2', 'width': 108}, {'height': 121, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=216&crop=smart&auto=webp&s=e9f6161a19b06707c892831a3194e5a6fd56820f', 'width': 216}, {'height': 180, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=320&crop=smart&auto=webp&s=07c58c72772ae9370aa24b1b1610a2a031e3ca05', 'width': 320}, {'height': 360, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=640&crop=smart&auto=webp&s=c8b81654823da10d969a3add2041b03f2485f5f7', 'width': 640}, {'height': 540, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=960&crop=smart&auto=webp&s=849589f60897a4c74979d0618fc3b8bd39ac380a', 'width': 960}, {'height': 607, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?width=1080&crop=smart&auto=webp&s=708e078c982e6960805ec22fef7ca3f01c46e42e', 'width': 1080}], 'source': {'height': 675, 'url': 'https://external-preview.redd.it/ZMOM1B0x6xcUYwapK2gRS5LZfZNqvYZAsJegcnmjQZs.jpg?auto=webp&s=0baebdc7a2287f9cc0f0b17ed79949b3c730ecd6', 'width': 1200}, 'variants': {}}]}
Alternative to llama.cpp?
0
Is there any other way to run GGUF than llama.cpp? I need some binaries that I can bundle up with an application, and llama.cpp for some reason ramdonly slows down quite a lot. Ollama is good, but for windows it kinda runs like a standalone application.
2025-01-21T10:40:06
https://www.reddit.com/r/LocalLLaMA/comments/1i6fu6g/alternative_to_llamacpp/
Few_Acanthisitta_858
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6fu6g
false
null
t3_1i6fu6g
/r/LocalLLaMA/comments/1i6fu6g/alternative_to_llamacpp/
false
false
self
0
null
Literally unusable
112
2025-01-21T10:47:42
https://i.redd.it/iatgsah1ubee1.png
WarlaxZ
i.redd.it
1970-01-01T00:00:00
0
{}
1i6fxxy
false
null
t3_1i6fxxy
/r/LocalLLaMA/comments/1i6fxxy/literally_unusable/
false
false
https://a.thumbs.redditm…cuT4inlbGg54.jpg
112
{'enabled': True, 'images': [{'id': '8lihdmpAluoU8T4xIYn8l_nqwF7p__RyucTFxkkth_g', 'resolutions': [{'height': 68, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?width=108&crop=smart&auto=webp&s=8826c7c277888cc366c029200fbc2bb1b870ba5f', 'width': 108}, {'height': 136, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?width=216&crop=smart&auto=webp&s=0e81777fa7d5ffeb0436617c8ed030d2b11ed03e', 'width': 216}, {'height': 202, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?width=320&crop=smart&auto=webp&s=033ee9e527bd5220d0d7eb70085206dcc8c4397d', 'width': 320}, {'height': 404, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?width=640&crop=smart&auto=webp&s=24ef211ed58771aa9136fe74e30af422f415bc31', 'width': 640}, {'height': 607, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?width=960&crop=smart&auto=webp&s=7a6806db7ac01ed98cb0f91cb349da6d33419823', 'width': 960}], 'source': {'height': 652, 'url': 'https://preview.redd.it/iatgsah1ubee1.png?auto=webp&s=29091bf60516aba66a1d12be9c58f2416fda4cfc', 'width': 1031}, 'variants': {}}]}
R1-Distill Tokenizer issue?
1
[removed]
2025-01-21T11:03:53
https://www.reddit.com/r/LocalLLaMA/comments/1i6g6df/r1distill_tokenizer_issue/
AdventurousSwim1312
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6g6df
false
null
t3_1i6g6df
/r/LocalLLaMA/comments/1i6g6df/r1distill_tokenizer_issue/
false
false
self
1
null
Code Completion in Enterprise
1
[removed]
2025-01-21T11:05:01
https://www.reddit.com/r/LocalLLaMA/comments/1i6g6yv/code_completion_in_enterprise/
CarelessPerformer571
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6g6yv
false
null
t3_1i6g6yv
/r/LocalLLaMA/comments/1i6g6yv/code_completion_in_enterprise/
false
false
self
1
null
Quick tutorial: How to run Cursor at home (R1-14b + Continue.dev)
17
Hey everyone! Quick post today - just thought I'd throw together a mini tutorial so more people can start using R1's awesome coding power for free at home 😁 The set-up I show here is running on my Mac, but it should be simple to run on pretty much any machine if you swap from using an MLX model to a GGUF model 😎 https://reddit.com/link/1i6g7zd/video/9l6kouajvbee1/player
2025-01-21T11:07:05
https://www.reddit.com/r/LocalLLaMA/comments/1i6g7zd/quick_tutorial_how_to_run_cursor_at_home_r114b/
mark-lord
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6g7zd
false
null
t3_1i6g7zd
/r/LocalLLaMA/comments/1i6g7zd/quick_tutorial_how_to_run_cursor_at_home_r114b/
false
false
self
17
null
Got DeepSeek R1 running locally - Full setup guide and my personal review (Free OpenAI o1 alternative that runs locally??)
8
Just discovered DeepSeek R1 and I'm pretty hyped about it. For those who don't know, it's a new **open-source AI model that matches OpenAI o1 and Claude 3.5 Sonnet** in math, coding, and reasoning tasks. You can check out Reddit to see what others are saying about DeepSeek R1 vs OpenAI o1 and Claude 3.5 Sonnet. For me it's really good - good enough to be compared with those top models. And the best part? **You can run it locally on your machine, with total privacy and 100% FREE!!** I've got it running locally and have been playing with it for a while. Here's my setup - super easy to follow: *(Just a note: While I'm using a Mac,* ***this guide works exactly the same for Windows and Linux users****! 👌)* **1) Install Ollama** Quick intro to Ollama: It's a tool for running AI models locally on your machine. Grab it here: [https://ollama.com/download](https://ollama.com/download) https://preview.redd.it/vdmiiuw4vbee1.png?width=748&format=png&auto=webp&s=2e1efb91eee9cfd8c654ed3282154e92cbbcedad **2) Next, you'll need to pull and run the DeepSeek R1 model locally.** Ollama offers different model sizes - basically, bigger models = smarter AI, but need better GPU. Here's the lineup: || || |1.5B version (smallest)|ollama run deepseek-r1:1.5b| |8B version|ollama run deepseek-r1:8b| |14B version|ollama run deepseek-r1:14b| |32B version|ollama run deepseek-r1:32b| |70B version (biggest/smartest)|ollama run deepseek-r1:70b| Maybe start with a smaller model first to test the waters. Just open your terminal and run: `ollama run deepseek-r1:8b` Once it's pulled, the model will run locally on your machine. Simple as that! *Note: The bigger versions (like 32B and 70B) need some serious GPU power. Start small and work your way up based on your hardware!* https://preview.redd.it/uk32frykvbee1.png?width=966&format=png&auto=webp&s=df7a11a9b2c03e89b899b9aa3d9e1b62fd194197 **3) Set up Chatbox - a powerful client for AI models** Quick intro to Chatbox: a free, clean, and powerful desktop interface that works with most models. I started it as a side project for 2 years. It’s privacy-focused (all data stays local) and super easy to set up—no Docker or complicated steps. Download here: [https://chatboxai.app](https://chatboxai.app) In Chatbox, go to settings and switch the model provider to Ollama. Since you're running models locally, you can ignore the built-in cloud AI options - **no license key or payment is needed!** https://preview.redd.it/ye2tfudmvbee1.png?width=1940&format=png&auto=webp&s=2711854eb585e6940c8fa27fa0fdc6c0e656fd03 Then set up the Ollama API host - the default setting is [`http://127.0.0.1:11434`](http://127.0.0.1:11434), which should work right out of the box. That's it! Just pick the model and hit save. Now you're all set and ready to chat with your locally running Deepseek R1! 🚀 https://preview.redd.it/vizcc81pvbee1.png?width=2238&format=png&auto=webp&s=b80cb5066444203c85fd5d267b710e991df2381f Hope this helps! Let me know if you run into any issues. \--------------------- Here are a few tests I ran on my local DeepSeek R1 setup (loving Chatbox's **artifact preview** feature btw!) 👇 **Explain TCP:** https://preview.redd.it/dqa138svvbee1.png?width=2268&format=png&auto=webp&s=47c01c70f596a22e1c4cfb85878f2dd539a47824 Honestly, this looks pretty good, especially considering it's just an 8B model! **Make a Pac-Man game:** https://preview.redd.it/kt5b3vmxvbee1.png?width=1244&format=png&auto=webp&s=391246ecdb837b943b98c20795bc27ba0c2087b6 It looks great, but I couldn’t actually play it. I feel like there might be a few small bugs that could be fixed with some tweaking. (Just to clarify, this wasn’t done on the local model — my mac doesn’t have enough space for the largest deepseek R1 70b model, so I used the cloud model instead.) \--------------------- Honestly, I’ve seen a lot of overhyped posts about models here lately, so I was a bit skeptical going into this. But after testing DeepSeek R1 myself, I think it’s actually really solid. It’s not some magic replacement for OpenAI or Claude, but it’s **surprisingly capable** for something that runs locally. The fact that it’s free and works offline is a huge plus. What do you guys think? Curious to hear your honest thoughts.
2025-01-21T11:12:06
https://www.reddit.com/r/LocalLLaMA/comments/1i6gahy/got_deepseek_r1_running_locally_full_setup_guide/
sleepingbenb
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6gahy
false
null
t3_1i6gahy
/r/LocalLLaMA/comments/1i6gahy/got_deepseek_r1_running_locally_full_setup_guide/
false
false
https://b.thumbs.redditm…j4xhtiRLyWSs.jpg
8
{'enabled': False, 'images': [{'id': 'krjt_5uhqcaDfYjfO7lkezThehav9cAIRJgcK-OKAmM', 'resolutions': [{'height': 56, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=108&crop=smart&auto=webp&s=53486800d92d75b19d59502534fa9ba2785c14b0', 'width': 108}, {'height': 113, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=216&crop=smart&auto=webp&s=b6f8fe68f176c90b3c2634702ce0e240165c319a', 'width': 216}, {'height': 168, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=320&crop=smart&auto=webp&s=ba4a7df526b23a412363b0285eb9709218cd0a0b', 'width': 320}, {'height': 336, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=640&crop=smart&auto=webp&s=1b231518e5ed41e809cceeaa1c12bf32733c2345', 'width': 640}, {'height': 504, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=960&crop=smart&auto=webp&s=69bbae7110c0f929d0a3e6682fde693305633de7', 'width': 960}, {'height': 567, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?width=1080&crop=smart&auto=webp&s=18433bdabee79410303b82563a6f388835945bef', 'width': 1080}], 'source': {'height': 630, 'url': 'https://external-preview.redd.it/s0D7i4Rco0trWh9Bu1uEkgnoJJLA3UNKUA9vs57seII.jpg?auto=webp&s=7a93b120137c378d21e25e2652789f870d1591a2', 'width': 1200}, 'variants': {}}]}
DeepSeek R1 32B on 2xA770s
4
Maximum seems to be about 65K of context without spilling into system RAM... But even when it does there is some leeway before getting painfully slow. Will publish more detailed benchmarks soon, but this seems to be one of the cheapest way to run and fine-tune larger LLMs ATM. I am also testing setups with 3 and 4 GPU and currently working on a multi-host solution with DeepSeed, where we hopefully get the 600B model to run!
2025-01-21T11:17:26
https://v.redd.it/ycrj91bezbee1
Ragecommie
v.redd.it
1970-01-01T00:00:00
0
{}
1i6gdb9
false
{'reddit_video': {'bitrate_kbps': 5000, 'dash_url': 'https://v.redd.it/ycrj91bezbee1/DASHPlaylist.mpd?a=1740050259%2CZGE0MjRiMGJhMDNlNjUxYjA3YzZmZTdmOWY1NDk3Y2U2N2RkNmFhMDZkZGRhYjNjZTI5ZTExOTQ3NTNiM2Y5ZQ%3D%3D&v=1&f=sd', 'duration': 18, 'fallback_url': 'https://v.redd.it/ycrj91bezbee1/DASH_1080.mp4?source=fallback', 'has_audio': True, 'height': 936, 'hls_url': 'https://v.redd.it/ycrj91bezbee1/HLSPlaylist.m3u8?a=1740050259%2CNzAxNmFhMWI0OGFjN2Q1YWMwZjRhMGExNTZmOGY4YTJkZGJiMDAyN2Y3MmMwYmQ3ZjBjOGE5NDA4NzdjYmNhZg%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/ycrj91bezbee1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 1920}}
t3_1i6gdb9
/r/LocalLLaMA/comments/1i6gdb9/deepseek_r1_32b_on_2xa770s/
false
false
https://external-preview…b908846f0b29d78d
4
{'enabled': False, 'images': [{'id': 'dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy', 'resolutions': [{'height': 52, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=108&crop=smart&format=pjpg&auto=webp&s=e914ade0ca832a88e55523309720f74f45e1ec44', 'width': 108}, {'height': 105, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=216&crop=smart&format=pjpg&auto=webp&s=103031e7afe6321f3369adc79c941beee181ec71', 'width': 216}, {'height': 156, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=320&crop=smart&format=pjpg&auto=webp&s=7ec90eb0defd261f64adc03a1de1f93c3008adc4', 'width': 320}, {'height': 312, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=640&crop=smart&format=pjpg&auto=webp&s=c24d6082b4bb2c3461d036ad0483cb47cbfc9f44', 'width': 640}, {'height': 468, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=960&crop=smart&format=pjpg&auto=webp&s=80abc00698627c1343eefc3b515cf3ea227ae388', 'width': 960}, {'height': 526, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?width=1080&crop=smart&format=pjpg&auto=webp&s=c82e10c1c2c4d0646e4da12f480175de57b22d04', 'width': 1080}], 'source': {'height': 1362, 'url': 'https://external-preview.redd.it/dTlqMmtuNGV6YmVlMTtRIh6-hKSvEvz63e23qY0vMLGcHYzT2stmVI0UPLDy.png?format=pjpg&auto=webp&s=f1c7396f67a26fb59496dfcba4caabb589f6657d', 'width': 2792}, 'variants': {}}]}
Abstract Multidimensional Structured Reasoning: Glyph Code Prompting
1
Alright everyone, just let me cook for a minute and then let me know if I am going crazy or if this is a useful thread to pull... To get straight to the point, I think I uncovered a new and potentially better way to not only prompt engineer LLMs but also improve their ability to reason in a dynamic yet structured way. All by harnessing In-Context Learning and providing the LLM with a more natural, intuitive tools set for itself. Here is an example of a small one-shot reasoning prompt: Execute this traversal, logic flow, synthesis, and generation process step by step using the provided context and logic in the following glyph code prompt: Abstract Tree of Thought Reasoning Thread-Flow {⦶("Abstract Symbolic Reasoning": "Dynamic Multidimensional Transformation and Extrapolation") ⟡("Objective": "Decode a sequence of evolving abstract symbols with multiple, interacting attributes and predict the next symbol in the sequence, along with a novel property not yet exhibited.") ⟡("Method": "Glyph-Guided Exploratory Reasoning and Inductive Inference") ⟡("Constraints": ω="High", ⋔="Hidden Multidimensional Rules, Non-Linear Transformations, Emergent Properties", "One-Shot Learning") ⥁{ (⊜⟡("Symbol Sequence": ⋔=" 1. ◇ (Vertical, Red, Solid) -> 2. ⬟ (Horizontal, Blue, Striped) -> 3. ○ (Vertical, Green, Solid) -> 4. ▴ (Horizontal, Red, Dotted) -> 5. ? ") -> ∿⟡("Initial Pattern Exploration": ⋔="Shape, Orientation, Color, Pattern")) ∿⟡("Initial Pattern Exploration") -> ⧓⟡("Attribute Clusters": ⋔="Geometric Transformations, Color Cycling, Pattern Alternation, Positional Relationships") ⧓⟡("Attribute Clusters") -> ⥁[ ⧓⟡("Branch": ⋔="Shape Transformation Logic") -> ∿⟡("Exploration": ⋔="Cyclic Sequence, Geometric Relationships, Symmetries"), ⧓⟡("Branch": ⋔="Orientation Dynamics") -> ∿⟡("Exploration": ⋔="Rotational Patterns, Axis Shifts, Inversion Rules"), ⧓⟡("Branch": ⋔="Color and Pattern Interaction") -> ∿⟡("Exploration": ⋔="Cyclic Permutations, Conditional Dependencies, Coupled Transformations"), ⧓⟡("Branch": ⋔="Positional Relationships") -> ∿⟡("Exploration": ⋔="Relative Movement, Spatial Constraints, Contextual Influence"), ⧓⟡("Branch": ⋔="Emergent Property Prediction") -> ∿⟡("Exploration": ⋔="Novel Attribute Introduction, Rule Extrapolation, Abstract Inference") ] ⥁(∿⟡("Exploration") -> ↑⟡("Hypotheses": ⋔="Candidate Rules for Each Attribute, Potential Interactions, Predicted Outcomes")) ↑⟡("Hypotheses") -> ⦑⟡("Integrated Model": ⋔="Combining Rules, Resolving Conflicts, Constructing a Unified Framework") ⦑⟡("Integrated Model") -> ✧⟡("Prediction": ⋔=" Fifth Symbol: - Shape: ? - Orientation: ? - Color: ? - Pattern: ? - Novel Property: ? (e.g., Size, Shading, Movement) Justification: ? (Explain the logical basis for each attribute prediction, referencing the discovered rules and their interactions.) ") } @Output(Prediction, Justification) @Reflect(Reasoning Process, Challenges, Insights, Comparison to Typical Reasoning Prompt Methods) @Engage now with full glyph code prompting logic, processing, and human-AI integrated interaction. } I know, that looks like a bunch of madness, but I am beginning to believe this allows the LLMs better access to more pertaining patterns and the ability to unpack the outputs within, leading to more specific, creative, and nuanced generations. I think this is the reason why libraries like SynthLang are so mysteriously powerful (https://github.com/ruvnet/SynthLang) For the logic and underlying hypothesis that governs all of this stuff, here is the most concise way I've been able to convey it. A longform post can be found at this link if you're curious (https://huggingface.co/blog/Severian/computational-model-for-symbolic-representations): The Computational Model for Symbolic Representations Framework introduces a method for enhancing human-AI collaboration by assigning user-defined symbolic representations (glyphs) to guide interactions with computational models. This interaction and syntax is called Glyph Code Prompting. Glyphs function as conceptual tags or anchors, representing abstract ideas, storytelling elements, or domains of focus (e.g., pacing, character development, thematic resonance). Users can steer the AI’s focus within specific conceptual domains by using these symbols, creating a shared framework for dynamic collaboration. Glyphs do not alter the underlying architecture of the AI; instead, they leverage and give new meaning to existing mechanisms such as contextual priming, attention mechanisms, and latent space activation within neural networks. This approach does not invent new capabilities within the AI but repurposes existing features. Neural networks are inherently designed to process context, prioritize input, and retrieve related patterns from their latent space. Glyphs build on these foundational capabilities, acting as overlays of symbolic meaning that channel the AI's probabilistic processes into specific focus areas. For example, consider the concept of 'trees'. In a typical LLM, this word might evoke a range of associations: biological data, environmental concerns, poetic imagery, or even data structures in computer science. Now, imagine a glyph, let's say `⟡`, when specifically defined to represent the vector cluster we will call "Arboreal Nexus". When used in a prompt, `⟡` would direct the model to emphasize dimensions tied to a complex, holistic understanding of trees that goes beyond a simple dictionary definition, pulling the latent space exploration into areas that include their symbolic meaning in literature and mythology, the scientific intricacies of their ecological roles, and the complex emotions they evoke in humans (such as longevity, resilience, and interconnectedness). Instead of a generic response about trees, the LLM, guided by `⟡` as defined in this instance, would generate text that reflects this deeper, more nuanced understanding of the concept: "Arboreal Nexus." This framework allows users to draw out richer, more intentional responses without modifying the underlying system by assigning this rich symbolic meaning to patterns already embedded within the AI's training data. **The Core Point: Glyphs, acting as collaboratively defined symbols linking related concepts, add a layer of multidimensional semantic richness to user-AI interactions by serving as contextual anchors that guide the AI's focus. This enhances the AI's ability to generate more nuanced and contextually appropriate responses. For instance, a symbol like** `!` **can carry multidimensional semantic meaning and connections, demonstrating the practical value of glyphs in conveying complex intentions efficiently.** Final Note: Please test this out and see what your experience is like. I am hoping to open up a discussion and see if any of this can be invalidated or validated. I haven't made a repo yet because I want to make sure I'm not crazy before going all in and pursuing this fully lol
2025-01-21T11:22:16
https://www.reddit.com/r/LocalLLaMA/comments/1i6gfs4/abstract_multidimensional_structured_reasoning/
vesudeva
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6gfs4
false
null
t3_1i6gfs4
/r/LocalLLaMA/comments/1i6gfs4/abstract_multidimensional_structured_reasoning/
false
false
self
1
{'enabled': False, 'images': [{'id': 'Y7-Rpq6fv-Ac-JFbxCCNdlFHc5Yp92rUCDinRjAVBpQ', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=108&crop=smart&auto=webp&s=4758932c6780a448abc5ba640b86c98af810dcee', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=216&crop=smart&auto=webp&s=3e4fc7d92a94b83376e1f0259669a66c42149e3c', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=320&crop=smart&auto=webp&s=c29f28fb2c936c12279f02c5017a96a452d4383e', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=640&crop=smart&auto=webp&s=3dbabf00df957ef225a7a91a2b4c99b30cf710ee', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=960&crop=smart&auto=webp&s=d0f8f3859aa49cf8e249a907c38c7d3824480ab7', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?width=1080&crop=smart&auto=webp&s=b3286bfa823e96288f5ed2e2dc44cbae58a7ff93', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/iAcUP5Oh_bYSr3_zupBo2q85J5aTZaEZwlLqR2I_ENk.jpg?auto=webp&s=4caf16ba1e7315630b28491fcc6add1698f7d0ea', 'width': 1200}, 'variants': {}}]}
I hooked DeepSeek-V3 to my thermal printer and created the first prompt-to-print tool
1
2025-01-21T11:22:35
https://i.redd.it/p9tgstdazbee1.jpeg
Printercow
i.redd.it
1970-01-01T00:00:00
0
{}
1i6gfxx
false
null
t3_1i6gfxx
/r/LocalLLaMA/comments/1i6gfxx/i_hooked_deepseekv3_to_my_thermal_printer_and/
false
false
https://b.thumbs.redditm…C-uVWWItwDuQ.jpg
1
{'enabled': True, 'images': [{'id': 'Uw5SF8TNkh7T84mM1Mpr8dTdzxCEzm7GXCU5cc8OrWA', 'resolutions': [{'height': 143, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?width=108&crop=smart&auto=webp&s=030ff4d71f86376a49b9cf279aeea8a797a7d4be', 'width': 108}, {'height': 287, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?width=216&crop=smart&auto=webp&s=402457547c72c749ec111c63306c9bf67902ca2e', 'width': 216}, {'height': 425, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?width=320&crop=smart&auto=webp&s=80bbe1bc478fab4ba64d6bed81c3c2d7f6f77e5b', 'width': 320}, {'height': 850, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?width=640&crop=smart&auto=webp&s=509354bbc19373f2cc278b86528077916480d42e', 'width': 640}, {'height': 1276, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?width=960&crop=smart&auto=webp&s=e351d74ce8eef618c33d8be414d5e2acb2f4de80', 'width': 960}], 'source': {'height': 1280, 'url': 'https://preview.redd.it/p9tgstdazbee1.jpeg?auto=webp&s=99907dc1f183629260c847ed35f2ca65800feac4', 'width': 963}, 'variants': {}}]}
Before I download 700gb of Deepseek R1
1
So I've seen many people test smaller distilled/quantized version of Deepseek on their macbook etc... Reading the required specs, I think I should be able to run the raw 700gb version on my machine but wondered if anyone did it. If you think I cant, whats the biggest version i could pull? Intel i9 13th - 64gb DDR5 6400 - 2Tb dedicated NVMe 7000mb/s - Nvidia 4090
2025-01-21T11:50:22
https://www.reddit.com/r/LocalLLaMA/comments/1i6gusy/before_i_download_700gb_of_deepseek_r1/
Accomplished_Law5807
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6gusy
false
null
t3_1i6gusy
/r/LocalLLaMA/comments/1i6gusy/before_i_download_700gb_of_deepseek_r1/
false
false
self
1
null
I hooked DeepSeek-V3 to my thermal printer and can print everything I want
1
2025-01-21T11:55:47
https://i.redd.it/fpk9d37j1cee1.jpeg
National_Strain_8384
i.redd.it
1970-01-01T00:00:00
0
{}
1i6gxsz
false
null
t3_1i6gxsz
/r/LocalLLaMA/comments/1i6gxsz/i_hooked_deepseekv3_to_my_thermal_printer_and_can/
false
false
https://b.thumbs.redditm…ewABiXE7zaVs.jpg
1
{'enabled': True, 'images': [{'id': 'HwQdtTv5oGSbn5OdPynG8-dDmyN6UKiPQlvIP_41t6s', 'resolutions': [{'height': 143, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?width=108&crop=smart&auto=webp&s=0547f0afa7ca25e4cebfaa6620a4c95161571c75', 'width': 108}, {'height': 287, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?width=216&crop=smart&auto=webp&s=7b1ffdc3eea0b8085ced3fc25ca7958b31dc8963', 'width': 216}, {'height': 425, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?width=320&crop=smart&auto=webp&s=38b754351d6ef60bd9caa78d7ef8a6de9ca1198d', 'width': 320}, {'height': 850, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?width=640&crop=smart&auto=webp&s=746da39535dfe83f7cde289268cf1e4486c47f7d', 'width': 640}, {'height': 1276, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?width=960&crop=smart&auto=webp&s=c8d51e2c9960b1edaa0bf91637ae2340f5e74383', 'width': 960}], 'source': {'height': 1280, 'url': 'https://preview.redd.it/fpk9d37j1cee1.jpeg?auto=webp&s=69d1d072b275bc0f62e73ffcde80a1ecad619909', 'width': 963}, 'variants': {}}]}
Free Deepseek Web Chat using R1?
1
Has the free DeepSeek web chat been updated to use the new R1 model instead of the R1-lite-preview? If yes, does this mean everyone now has basically unlimited access to an O1-level model easily?
2025-01-21T12:15:19
https://www.reddit.com/r/LocalLLaMA/comments/1i6h9kd/free_deepseek_web_chat_using_r1/
jpgirardi
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6h9kd
false
null
t3_1i6h9kd
/r/LocalLLaMA/comments/1i6h9kd/free_deepseek_web_chat_using_r1/
false
false
self
1
null
Y’all this hype train is so real (im sure ill get deleted lol)
0
Pics and explanation to follow as comments! This is talking about Deepseek R1’s-7B-Distill models and *holy shiznit* lmao, like danggggggggggg. I don’t have very much luck posting though so hopefully this post survives lol (watch it be too long…)
2025-01-21T12:22:55
https://v.redd.it/85f5bmq1bcee1
clduab11
/r/LocalLLaMA/comments/1i6hdz0/yall_this_hype_train_is_so_real_im_sure_ill_get/
1970-01-01T00:00:00
0
{}
1i6hdz0
false
{'reddit_video': {'bitrate_kbps': 1200, 'dash_url': 'https://v.redd.it/85f5bmq1bcee1/DASHPlaylist.mpd?a=1740183784%2CMDFjZDZhNmRkY2QwOTYwOTEyYzY4ZjFkZGMzOGJkMTFhMjFlYThmNmIwNzYyMzY5NWZkNjI5YjNiY2JiODk4Mg%3D%3D&v=1&f=sd', 'duration': 102, 'fallback_url': 'https://v.redd.it/85f5bmq1bcee1/DASH_480.mp4?source=fallback', 'has_audio': True, 'height': 338, 'hls_url': 'https://v.redd.it/85f5bmq1bcee1/HLSPlaylist.m3u8?a=1740183784%2CZDhlODIzMjEwNWVmOTI0ZmY0ZWVjOWIzODU2YTgwYzAxMjA4N2E3YzhiYjViMzM1NzE0NzJjN2U2NDg3Mzc0ZA%3D%3D&v=1&f=sd', 'is_gif': False, 'scrubber_media_url': 'https://v.redd.it/85f5bmq1bcee1/DASH_96.mp4', 'transcoding_status': 'completed', 'width': 854}}
t3_1i6hdz0
/r/LocalLLaMA/comments/1i6hdz0/yall_this_hype_train_is_so_real_im_sure_ill_get/
false
false
https://external-preview…948eb067972bbe3c
0
{'enabled': False, 'images': [{'id': 'MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT', 'resolutions': [{'height': 42, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=108&crop=smart&format=pjpg&auto=webp&s=7382ab94ba30e1882a0267ef2b82484e7c2e8c66', 'width': 108}, {'height': 85, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=216&crop=smart&format=pjpg&auto=webp&s=adf156bf0b4bd6982d537ac98593803e73b27813', 'width': 216}, {'height': 126, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=320&crop=smart&format=pjpg&auto=webp&s=9d044efb92da3d4db126387119ebd1c8ea386ef7', 'width': 320}, {'height': 252, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=640&crop=smart&format=pjpg&auto=webp&s=b75a1590103b5e4baec68e6b29e1afcefd6aa07d', 'width': 640}, {'height': 379, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=960&crop=smart&format=pjpg&auto=webp&s=afadade7401e31d1bb20f0a94126c05412c76eb1', 'width': 960}, {'height': 426, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?width=1080&crop=smart&format=pjpg&auto=webp&s=62678f2bc31105996cd10d789886fadcda47b241', 'width': 1080}], 'source': {'height': 636, 'url': 'https://external-preview.redd.it/MGFsM21zbzFiY2VlMf66U_pM71zptNgFvyQDl9pw39P_DcoqNmfaVXQjXgJT.png?format=pjpg&auto=webp&s=5adeefbbca9f75ef216a217d42a4ca4eaaac1f5e', 'width': 1610}, 'variants': {}}]}
Model Searching
1
[removed]
2025-01-21T12:39:29
https://www.reddit.com/r/LocalLLaMA/comments/1i6hnzt/model_searching/
Arcade_Gamer21
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6hnzt
false
null
t3_1i6hnzt
/r/LocalLLaMA/comments/1i6hnzt/model_searching/
false
false
self
1
null
How I Use Local AI to Automatically Catalog My Game Collection from Pictures
1
2025-01-21T12:46:46
https://medium.com/@rafaelbenari/how-i-automatically-catalog-my-game-collection-from-pictures-79416ae4b32c
Oatilis
medium.com
1970-01-01T00:00:00
0
{}
1i6hssy
false
null
t3_1i6hssy
/r/LocalLLaMA/comments/1i6hssy/how_i_use_local_ai_to_automatically_catalog_my/
false
false
https://b.thumbs.redditm…tzGsru4R5TtE.jpg
1
{'enabled': False, 'images': [{'id': 'RIj26ANlQShGWBx1vIqv_RHZVnjAo75ojLXTFoKseSM', 'resolutions': [{'height': 63, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=108&crop=smart&auto=webp&s=c320d07c8eccab4c7247ae430b51199e91059009', 'width': 108}, {'height': 126, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=216&crop=smart&auto=webp&s=9a213e11e78b42b68bb43bf1ceae191c06da74f8', 'width': 216}, {'height': 186, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=320&crop=smart&auto=webp&s=3725f3405fb4bca479e1991f997960781f79d645', 'width': 320}, {'height': 373, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=640&crop=smart&auto=webp&s=e883294e0996e912c8f44d6b43db2750c6192983', 'width': 640}, {'height': 560, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=960&crop=smart&auto=webp&s=d819db490de3ca81e97bfa025038f7d722d8a366', 'width': 960}, {'height': 630, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=1080&crop=smart&auto=webp&s=d8712e298c38a8e46dd16d7de1947253d96f841d', 'width': 1080}], 'source': {'height': 700, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?auto=webp&s=082138db20a079736b10a88c7b64fd566a66b048', 'width': 1200}, 'variants': {}}]}
Need advice: Sub-20k€ Deep Learning Build - 5090 vs ADA 6000?
1
[removed]
2025-01-21T12:51:12
https://www.reddit.com/r/LocalLLaMA/comments/1i6hvjv/need_advice_sub20k_deep_learning_build_5090_vs/
StudentOfChaos123
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6hvjv
false
null
t3_1i6hvjv
/r/LocalLLaMA/comments/1i6hvjv/need_advice_sub20k_deep_learning_build_5090_vs/
false
false
self
1
null
Hi need some help I'm noob here
2
Hi am new to this local llm thing I want to run an llm on my laptop which has the following specks I5 11300h 8gb ram GTX 1650 I'm using lm studio to run llm. I want this llm to be good in programming with story writing. I know both are pollar opposite but that my requirement. Thanks for reading my post
2025-01-21T12:52:55
https://www.reddit.com/r/LocalLLaMA/comments/1i6hwnb/hi_need_some_help_im_noob_here/
Aware-Fudge-6146
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6hwnb
false
null
t3_1i6hwnb
/r/LocalLLaMA/comments/1i6hwnb/hi_need_some_help_im_noob_here/
false
false
self
2
null
Actual video of DeepSeek-R1 being trained.
16
2025-01-21T12:54:09
https://i.redd.it/wl8i1xsngcee1.gif
Porespellar
i.redd.it
1970-01-01T00:00:00
0
{}
1i6hxfi
false
null
t3_1i6hxfi
/r/LocalLLaMA/comments/1i6hxfi/actual_video_of_deepseekr1_being_trained/
false
false
https://b.thumbs.redditm…ZaeR7XQiyC9I.jpg
16
{'enabled': True, 'images': [{'id': 'hyN-9ZnFc5ogkvi-oNQ3hmaC6d8y7tPnsPKGziFw4Mo', 'resolutions': [{'height': 81, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=108&crop=smart&format=png8&s=73919b35d5d30ce21ac0332ef261d1fad0a53ea3', 'width': 108}, {'height': 162, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=216&crop=smart&format=png8&s=76b050f182a416742d19b961d475224e68861937', 'width': 216}], 'source': {'height': 165, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?format=png8&s=625f34b258e702c3e82dea477200330f795b6aaa', 'width': 220}, 'variants': {'gif': {'resolutions': [{'height': 81, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=108&crop=smart&s=515c18fb20f47ed2cc0a607ea26f060f75556ea1', 'width': 108}, {'height': 162, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=216&crop=smart&s=dcefb831252f3b3704759ddddc2a5852dd775ba7', 'width': 216}], 'source': {'height': 165, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?s=732da4b14d37734dc8b982a3729f90ce99128897', 'width': 220}}, 'mp4': {'resolutions': [{'height': 81, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=108&format=mp4&s=2e108ea62ec455a11b83d654641b82626f7646f6', 'width': 108}, {'height': 162, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?width=216&format=mp4&s=2feeddcbd10bf96028238cd7818f182ffbb4c866', 'width': 216}], 'source': {'height': 165, 'url': 'https://preview.redd.it/wl8i1xsngcee1.gif?format=mp4&s=63793f459f66eb910ea60289e1843b962cda729c', 'width': 220}}}}]}
Hugging Face links expire now?
3
I set deepseek 70b to download over night and I woke up to a ton of 403 forbidden messages instead of the model. Now this morning I queued up a 6bit quant and after 2 files the links say 403 again. How are we internet challenged people supposed to download models over night with these changes? Is this only happening to me perchance?
2025-01-21T12:57:21
https://www.reddit.com/r/LocalLLaMA/comments/1i6hzil/hugging_face_links_expire_now/
a_beautiful_rhind
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6hzil
false
null
t3_1i6hzil
/r/LocalLLaMA/comments/1i6hzil/hugging_face_links_expire_now/
false
false
self
3
{'enabled': False, 'images': [{'id': 'GGvwJGmkoijQC6DxfLJh4EaQOyRtQ07ppUyTl7SZfYk', 'resolutions': [{'height': 48, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=108&crop=smart&auto=webp&s=bd444b1610df902c4f6508f93ff426abffa80627', 'width': 108}, {'height': 96, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=216&crop=smart&auto=webp&s=f9e4e0d5cf5fa320885ee9af174add379d5b615f', 'width': 216}, {'height': 143, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=320&crop=smart&auto=webp&s=d933563ced52541148573a216dba762c2301be03', 'width': 320}, {'height': 287, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=640&crop=smart&auto=webp&s=0fabed7fffe0084325f93c91f04de377074b2f11', 'width': 640}, {'height': 430, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=960&crop=smart&auto=webp&s=3f9a9462c689a92f74d17575013408f3a8c2ed16', 'width': 960}, {'height': 484, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?width=1080&crop=smart&auto=webp&s=6141bde5b024b7fcfc5080a78dc27a2349f7961d', 'width': 1080}], 'source': {'height': 608, 'url': 'https://external-preview.redd.it/qip6LB9LdJZUgktK7YOBoVYF3W1TbUgi4gdl5feOpNI.png?auto=webp&s=baa35aab6812aa88ea5773569a5bd6594bfedd13', 'width': 1355}, 'variants': {}}]}
is there a portable version of f5-tts? or know how to create one?
1
[removed]
2025-01-21T13:03:59
https://www.reddit.com/r/LocalLLaMA/comments/1i6i40c/is_there_a_portable_version_of_f5tts_or_know_how/
Wonderful-Fudge-5880
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6i40c
false
null
t3_1i6i40c
/r/LocalLLaMA/comments/1i6i40c/is_there_a_portable_version_of_f5tts_or_know_how/
false
false
self
1
{'enabled': False, 'images': [{'id': 'u2VOVFVnnAXjEuidlHshNowChxNJKliPwjS-lgpgtdo', 'resolutions': [{'height': 54, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=108&crop=smart&auto=webp&s=31c0261ceae9124c24fbae8c769293c835bad2e7', 'width': 108}, {'height': 108, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=216&crop=smart&auto=webp&s=04d61a95a5058d5246dbb0cb4e0439864fc4e778', 'width': 216}, {'height': 160, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=320&crop=smart&auto=webp&s=5bf720057c76d000085c7e712f4766fa251f70e4', 'width': 320}, {'height': 320, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=640&crop=smart&auto=webp&s=3faeba095eadd126f6826c262386f54b801d2294', 'width': 640}, {'height': 480, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=960&crop=smart&auto=webp&s=683cf52b02eecccd872a0f7e0f6ce887b3348208', 'width': 960}, {'height': 540, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?width=1080&crop=smart&auto=webp&s=5cbff096aed86195b42095e8b15de5f6d0321f45', 'width': 1080}], 'source': {'height': 600, 'url': 'https://external-preview.redd.it/4zZqhe3SmxJi9IWadPeeYqGVwd2sQ1RRlX-gOG35Zro.jpg?auto=webp&s=d96845c61ca69ee9617cfc616c3d8d94211eca32', 'width': 1200}, 'variants': {}}]}
Using Qwen2-VL Locally to Automatically Catalog My Game Collection from Pictures
1
2025-01-21T13:13:41
https://medium.com/@rafaelbenari/how-i-automatically-catalog-my-game-collection-from-pictures-79416ae4b32c
Oatilis
medium.com
1970-01-01T00:00:00
0
{}
1i6iaij
false
null
t3_1i6iaij
/r/LocalLLaMA/comments/1i6iaij/using_qwen2vl_locally_to_automatically_catalog_my/
false
false
https://b.thumbs.redditm…tzGsru4R5TtE.jpg
1
{'enabled': False, 'images': [{'id': 'RIj26ANlQShGWBx1vIqv_RHZVnjAo75ojLXTFoKseSM', 'resolutions': [{'height': 63, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=108&crop=smart&auto=webp&s=c320d07c8eccab4c7247ae430b51199e91059009', 'width': 108}, {'height': 126, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=216&crop=smart&auto=webp&s=9a213e11e78b42b68bb43bf1ceae191c06da74f8', 'width': 216}, {'height': 186, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=320&crop=smart&auto=webp&s=3725f3405fb4bca479e1991f997960781f79d645', 'width': 320}, {'height': 373, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=640&crop=smart&auto=webp&s=e883294e0996e912c8f44d6b43db2750c6192983', 'width': 640}, {'height': 560, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=960&crop=smart&auto=webp&s=d819db490de3ca81e97bfa025038f7d722d8a366', 'width': 960}, {'height': 630, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?width=1080&crop=smart&auto=webp&s=d8712e298c38a8e46dd16d7de1947253d96f841d', 'width': 1080}], 'source': {'height': 700, 'url': 'https://external-preview.redd.it/3rTGfND7o6ypgTxpPQyFu4E8VLgC7_frn_C_n1aCPCU.jpg?auto=webp&s=082138db20a079736b10a88c7b64fd566a66b048', 'width': 1200}, 'variants': {}}]}
Need help in using a TTS model in Home Assistant
3
I'm a noob in AI, but I'm not a noob in self-hosting. I have finally found a Ukrainian TTS model that sounds not particularly bad: https://huggingface.co/spaces/patriotyk/styletts2-ukrainian Now I need to make it speak for me in my Home Assistant instance. I have Open-Webui, Piper, [Speaches](https://github.com/speaches-ai/speaches) with [Custom OpenAI TTS](https://github.com/sfortis/openai_tts), I can deploy any other service, but I have no idea where to start in making this model work for me :( Piper needs a .json file similar to [this](https://huggingface.co/rhasspy/piper-voices/blob/main/voices.json), but I have no JSON files for this model :( Halp
2025-01-21T13:40:38
https://www.reddit.com/r/LocalLLaMA/comments/1i6isvz/need_help_in_using_a_tts_model_in_home_assistant/
ALERTua
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6isvz
false
null
t3_1i6isvz
/r/LocalLLaMA/comments/1i6isvz/need_help_in_using_a_tts_model_in_home_assistant/
false
false
self
3
{'enabled': False, 'images': [{'id': 'Qo7kP8XDiX6nDZQLMMySZUEaQFZjRy9qwNjOqwdDWbc', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=108&crop=smart&auto=webp&s=1c8ea40e34aa64df71638f55e2b2f92d5a7526c8', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=216&crop=smart&auto=webp&s=f41d2ac845bcb14afe73453195e5d21c71a8019d', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=320&crop=smart&auto=webp&s=653192504a8462e484daedba07148910fc033c3b', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=640&crop=smart&auto=webp&s=fee1200ca26a195e4d4e7228eaa6dae321eb4653', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=960&crop=smart&auto=webp&s=41844a70fb7955c661896fa440f6e877bb656aa8', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?width=1080&crop=smart&auto=webp&s=f473cc2decd0e84807c679034967052f7918d23b', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/Lhly4f1dpTBTPiIxFhwkFDe5IxIhhf3heGGGAUO87HI.jpg?auto=webp&s=abd3a553696dfd1fc55296eeb83051dc416b9767', 'width': 1200}, 'variants': {}}]}
OpenAI funded the Frontier Math dataset and had a contract in place to withhold disclosure until the O3 launch. thoughts?
0
[https://www.lesswrong.com/posts/cu2E8wgmbdZbqeWqb/meemi-s-shortform](https://www.lesswrong.com/posts/cu2E8wgmbdZbqeWqb/meemi-s-shortform)
2025-01-21T13:51:04
https://www.reddit.com/r/LocalLLaMA/comments/1i6j060/openai_funded_the_frontier_math_dataset_and_had_a/
cypherend1
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6j060
false
null
t3_1i6j060
/r/LocalLLaMA/comments/1i6j060/openai_funded_the_frontier_math_dataset_and_had_a/
false
false
self
0
{'enabled': False, 'images': [{'id': 'vYvgH_kHPmVGq0OLpHkQ4f-YObUhzba8fZBmsSAWcjM', 'resolutions': [{'height': 45, 'url': 'https://external-preview.redd.it/_kGMR5OcBR1_XR7JtRgy0kPUTKSsEWDRWeHlxSX3T_w.jpg?width=108&crop=smart&auto=webp&s=13906d384579edd182766ac4a4f338c1f727a59a', 'width': 108}, {'height': 91, 'url': 'https://external-preview.redd.it/_kGMR5OcBR1_XR7JtRgy0kPUTKSsEWDRWeHlxSX3T_w.jpg?width=216&crop=smart&auto=webp&s=50345aa3e0745c3b0d7e9b31142f661fad2e3cdf', 'width': 216}, {'height': 135, 'url': 'https://external-preview.redd.it/_kGMR5OcBR1_XR7JtRgy0kPUTKSsEWDRWeHlxSX3T_w.jpg?width=320&crop=smart&auto=webp&s=2974be59a5c5d027d56b1a40883ae53cf5c1eec0', 'width': 320}, {'height': 271, 'url': 'https://external-preview.redd.it/_kGMR5OcBR1_XR7JtRgy0kPUTKSsEWDRWeHlxSX3T_w.jpg?width=640&crop=smart&auto=webp&s=fd11b484225f901936bd8d694d309b5c885f84f6', 'width': 640}], 'source': {'height': 295, 'url': 'https://external-preview.redd.it/_kGMR5OcBR1_XR7JtRgy0kPUTKSsEWDRWeHlxSX3T_w.jpg?auto=webp&s=5be1f255d60c4268fe22fbba92057753809048bd', 'width': 696}, 'variants': {}}]}
DeepSeek R1 (Qwen 32B Distill) is now available for free on HuggingChat!
464
2025-01-21T14:07:01
https://hf.co/chat/models/deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
SensitiveCranberry
hf.co
1970-01-01T00:00:00
0
{}
1i6jbur
false
null
t3_1i6jbur
/r/LocalLLaMA/comments/1i6jbur/deepseek_r1_qwen_32b_distill_is_now_available_for/
false
false
https://a.thumbs.redditm…Cwl5dwqjWQT8.jpg
464
{'enabled': False, 'images': [{'id': '6IOVFsF0Sat_juRGMVTRxUgdy1tjum3AdY9uGARuUx0', 'resolutions': [{'height': 58, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=108&crop=smart&auto=webp&s=f2c35f5a1853716f58d4e4a4a05212bcb2a43428', 'width': 108}, {'height': 116, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=216&crop=smart&auto=webp&s=4043c1d88632c76cd7c9c972d752d7be17b89581', 'width': 216}, {'height': 172, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=320&crop=smart&auto=webp&s=1b881645a173059f76868c50b0c56eb2bf019b9e', 'width': 320}, {'height': 345, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=640&crop=smart&auto=webp&s=32931922738bce5caa4226b968a442514cf96587', 'width': 640}, {'height': 518, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=960&crop=smart&auto=webp&s=2b63b4e694d3509fd2fa43cd8b8f3f13f61f650a', 'width': 960}, {'height': 583, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?width=1080&crop=smart&auto=webp&s=094676a14460b65c5505d6cc9b95c8d1b581a04d', 'width': 1080}], 'source': {'height': 648, 'url': 'https://external-preview.redd.it/8HwSiZPd8K_hder46_kXYrWF23xE0qYwa1myzoXXUfM.jpg?auto=webp&s=22fff5a82e82ff5c3f7ba6834d21ad4966f5afcf', 'width': 1200}, 'variants': {}}]}
I'm finding code produced by R1 to contain less fluff than Sonnet
26
I've only done a little testing so far, but this has been true throughout. This is only a comment on the style of code rather than the quality/functionality. The absence of comments for every little thing is noticeable. With Sonnet, it is like it includes a full beginning-to-end guide for the code it writes via comments. R1 doesn't seem to do this at all - using comments only for occasional clarity or to denote sections. Overall, it seems to all be less verbose. That feels like a small difference which is actually a pretty big win, since it means feeding it back into R1 uses fewer tokens. This has been a long-standing gripe of mine with Sonnet, and pretty much every model since 3.5. Less fluff is nice.
2025-01-21T14:09:23
https://www.reddit.com/r/LocalLLaMA/comments/1i6jdlp/im_finding_code_produced_by_r1_to_contain_less/
EmbarrassedBiscotti9
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6jdlp
false
null
t3_1i6jdlp
/r/LocalLLaMA/comments/1i6jdlp/im_finding_code_produced_by_r1_to_contain_less/
false
false
self
26
null
Cursor now has DeepSeek V3 support
12
https://preview.redd.it/…37ce6d36cf086a
2025-01-21T14:15:40
https://www.reddit.com/r/LocalLLaMA/comments/1i6ji9h/cursor_now_has_deepseek_v3_support/
raucousbasilisk
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6ji9h
false
null
t3_1i6ji9h
/r/LocalLLaMA/comments/1i6ji9h/cursor_now_has_deepseek_v3_support/
false
false
https://a.thumbs.redditm…D1p9l4vwc5N4.jpg
12
null
How do people fine tune LLMs to not answer specific questions?
1
[removed]
2025-01-21T14:21:05
https://www.reddit.com/r/LocalLLaMA/comments/1i6jmc7/how_do_people_fine_tune_llms_to_not_answer/
AdvertisingOk9066
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6jmc7
false
null
t3_1i6jmc7
/r/LocalLLaMA/comments/1i6jmc7/how_do_people_fine_tune_llms_to_not_answer/
false
false
self
1
null
Deepseek app - How does it do web search?
1
How does the deepseek app do web search? Does anyone know what services they might be using? Also, is it possible to use their search functionality through API? Unable to find this in their docs.
2025-01-21T14:29:15
https://www.reddit.com/r/LocalLLaMA/comments/1i6jsl1/deepseek_app_how_does_it_do_web_search/
ttbap
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6jsl1
false
null
t3_1i6jsl1
/r/LocalLLaMA/comments/1i6jsl1/deepseek_app_how_does_it_do_web_search/
false
false
self
1
null
How do people fine tune LLMs to not answer specific questions?
1
[removed]
2025-01-21T14:31:34
https://www.reddit.com/r/LocalLLaMA/comments/1i6jubn/how_do_people_fine_tune_llms_to_not_answer/
AdvertisingOk9066
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6jubn
false
null
t3_1i6jubn
/r/LocalLLaMA/comments/1i6jubn/how_do_people_fine_tune_llms_to_not_answer/
false
false
self
1
null
Anyone tried running LLamafile on the new OnePlus 13?
1
[removed]
2025-01-21T14:37:34
https://www.reddit.com/r/LocalLLaMA/comments/1i6jysj/anyone_tried_running_llamafile_on_the_new_oneplus/
TechGent79
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6jysj
false
null
t3_1i6jysj
/r/LocalLLaMA/comments/1i6jysj/anyone_tried_running_llamafile_on_the_new_oneplus/
false
false
self
1
null
Deepseek R1 Agents
5
R1 and distillates are very much hot off the press. So far we have seen merges, quants and some finetunes. But not much discussion about how well these models perform with text-to-text agentic tasks. Has anyone tested **existing** workflows with the distilled models? Particularly orchestration systems? While 32B is appealing smaller models in larger teams offer a richer value proposition than what was available even two weeks ago for most of the local crowd. Teams of small models, beyond just task adherence and all the other heuristics we use outside of benchmarks, might enable better hardware utilization than what can be gained from a large model just from value it can deliver. Maybe small reasoning models are better at querying against a manager model for "help"- perhaps R1 data was the secret sauce small models have needed to get better at recognizing when things are outside of it's capability i.e, does guessing to fulfill the request *really* satisfy the instructions, or is the better response to state that it doesn't know *without* instructions to behave that way? Getting small text agents working well could be a *huge* leap forward for inference at the edge. I cant be the only person thinking about this- what do you guys think?
2025-01-21T14:40:34
https://www.reddit.com/r/LocalLLaMA/comments/1i6k11x/deepseek_r1_agents/
Echo9Zulu-
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6k11x
false
null
t3_1i6k11x
/r/LocalLLaMA/comments/1i6k11x/deepseek_r1_agents/
false
false
self
5
null
New LLaMA model on lmarena?
11
I hit a model called "experimental-router-0112" on lmarena and asked "Who made you and what is your model name" 3 times. Every time, it told me it is a model made by Meta based on LLaMA. 2 of the 3 times, it took quite a long time to answer (\~12 seconds) and the other time it almost immediately answered which considering the name leads me to speculate it is a router picking between a very large or a reasoning model and a smaller model. I saw some people on reddit say it reasons well and may be o3-mini. What do you think?
2025-01-21T14:42:48
https://www.reddit.com/r/LocalLLaMA/comments/1i6k2s1/new_llama_model_on_lmarena/
heyhellousername
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6k2s1
false
null
t3_1i6k2s1
/r/LocalLLaMA/comments/1i6k2s1/new_llama_model_on_lmarena/
false
false
self
11
null
What the field look like right now
1
OpenAI: hype & news DeepSeek: exceptional technical report & truly open AI models
2025-01-21T14:50:41
https://www.reddit.com/r/LocalLLaMA/comments/1i6k8y5/what_the_field_look_like_right_now/
vinhnx
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6k8y5
false
null
t3_1i6k8y5
/r/LocalLLaMA/comments/1i6k8y5/what_the_field_look_like_right_now/
false
false
self
1
null
DeepSeek-R1-Distill-Llama-8B 4bit on MBP M1 Max is awesome
1
[removed]
2025-01-21T14:56:01
https://www.reddit.com/r/LocalLLaMA/comments/1i6kd87/deepseekr1distillllama8b_4bit_on_mbp_m1_max_is/
Valuable-Run2129
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6kd87
false
null
t3_1i6kd87
/r/LocalLLaMA/comments/1i6kd87/deepseekr1distillllama8b_4bit_on_mbp_m1_max_is/
false
false
self
1
null
Experienced People Only Plz: How is R1 set up properly in Ollama
1
[removed]
2025-01-21T15:03:47
https://www.reddit.com/r/LocalLLaMA/comments/1i6kjug/experienced_people_only_plz_how_is_r1_set_up/
Rollingsound514
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6kjug
false
null
t3_1i6kjug
/r/LocalLLaMA/comments/1i6kjug/experienced_people_only_plz_how_is_r1_set_up/
false
false
self
1
null
You can now use both R1 and Search Web (see the comparison with and without R1)
68
2025-01-21T15:16:32
https://www.reddit.com/gallery/1i6ku6i
mw11n19
reddit.com
1970-01-01T00:00:00
0
{}
1i6ku6i
false
null
t3_1i6ku6i
/r/LocalLLaMA/comments/1i6ku6i/you_can_now_use_both_r1_and_search_web_see_the/
false
false
https://b.thumbs.redditm…7IhtpMIuEe9A.jpg
68
null
just tell it to be logical
75
2025-01-21T15:19:52
https://i.redd.it/j6ax54qh6dee1.png
spirobel
i.redd.it
1970-01-01T00:00:00
0
{}
1i6kwy7
false
null
t3_1i6kwy7
/r/LocalLLaMA/comments/1i6kwy7/just_tell_it_to_be_logical/
false
false
https://b.thumbs.redditm…D2OldG_LDmaM.jpg
75
{'enabled': True, 'images': [{'id': 'Xh5WWC_D5vSeCiRid1CzI5cN_8iINUhdauldQC1Uu54', 'resolutions': [{'height': 82, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?width=108&crop=smart&auto=webp&s=6af6b35c4621871097619b7f6689f3cec14057b5', 'width': 108}, {'height': 165, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?width=216&crop=smart&auto=webp&s=be4bdae7e262e5da25302b32924923158e8803dc', 'width': 216}, {'height': 245, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?width=320&crop=smart&auto=webp&s=5220fe9f6f8e3422dca786bc073d49b9a4b09389', 'width': 320}, {'height': 490, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?width=640&crop=smart&auto=webp&s=9f2541b2247a45ef3e36854545cd09c866802e79', 'width': 640}, {'height': 736, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?width=960&crop=smart&auto=webp&s=0a45fda461ad101fbb6cfa5ac366e8ae0afddb42', 'width': 960}], 'source': {'height': 793, 'url': 'https://preview.redd.it/j6ax54qh6dee1.png?auto=webp&s=a82d563100f1d2b1817ccc8c2e198577a12eefe4', 'width': 1034}, 'variants': {}}]}
The Chinese OBLITERATED OpenAI. A side-by-side comparison of DeepSeek R1 vs OpenAI O1 for Finance
1
[removed]
2025-01-21T15:26:39
https://www.reddit.com/r/LocalLLaMA/comments/1i6l2g3/the_chinese_obliterated_openai_a_sidebyside/
TORNADOig
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6l2g3
false
null
t3_1i6l2g3
/r/LocalLLaMA/comments/1i6l2g3/the_chinese_obliterated_openai_a_sidebyside/
false
false
self
1
{'enabled': False, 'images': [{'id': 'dJ_o2U5ZXBL-ud5nz6CRzNzSBUZdQk3G8h_Gkc4uwvs', 'resolutions': [{'height': 84, 'url': 'https://external-preview.redd.it/-whPSOrfhZnjoAX-irc72c5zf-vgt9dywVvSXP-IMlI.png?width=108&crop=smart&auto=webp&s=35b11d9c83ae12d8d50b8a3c8ba4833314ed01fa', 'width': 108}, {'height': 169, 'url': 'https://external-preview.redd.it/-whPSOrfhZnjoAX-irc72c5zf-vgt9dywVvSXP-IMlI.png?width=216&crop=smart&auto=webp&s=72c5c40f1f027eaa87481afc125a933d1006623f', 'width': 216}, {'height': 251, 'url': 'https://external-preview.redd.it/-whPSOrfhZnjoAX-irc72c5zf-vgt9dywVvSXP-IMlI.png?width=320&crop=smart&auto=webp&s=a5b65fc3c82de067c2f11d249e60b504ac5a2cc7', 'width': 320}, {'height': 503, 'url': 'https://external-preview.redd.it/-whPSOrfhZnjoAX-irc72c5zf-vgt9dywVvSXP-IMlI.png?width=640&crop=smart&auto=webp&s=3fa64ef99af605f81888f47412e7324ebe8c9140', 'width': 640}], 'source': {'height': 560, 'url': 'https://external-preview.redd.it/-whPSOrfhZnjoAX-irc72c5zf-vgt9dywVvSXP-IMlI.png?auto=webp&s=77be400a702bcaa0829d0fcfd090fde71a90e388', 'width': 712}, 'variants': {}}]}
Thanks to DeepSeek other open model releases with "research" license will be laughable
199
Imagine labs like Mistral, Cohere (do you remember them?) release open-weight model with so called research purposes only license. Comedy Central would call them for movie rights ;)
2025-01-21T15:28:03
https://www.reddit.com/r/LocalLLaMA/comments/1i6l3ms/thanks_to_deepseek_other_open_model_releases_with/
robertpiosik
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6l3ms
false
null
t3_1i6l3ms
/r/LocalLLaMA/comments/1i6l3ms/thanks_to_deepseek_other_open_model_releases_with/
false
false
self
199
null
Would you share access to your Local LLM with other people ?
1
I am building a platform to allow people to share their local llm with other users on the internet. Would that be something you would be intreseted in? Here is the waiting list: [https://loco.framer.website/](https://loco.framer.website/)
2025-01-21T15:33:11
https://www.reddit.com/r/LocalLLaMA/comments/1i6l7sd/would_you_share_access_to_your_local_llm_with/
Sarcinismo
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6l7sd
false
null
t3_1i6l7sd
/r/LocalLLaMA/comments/1i6l7sd/would_you_share_access_to_your_local_llm_with/
false
false
self
1
null
Local llama run pod vscode setup?
1
[removed]
2025-01-21T15:42:32
https://www.reddit.com/r/LocalLLaMA/comments/1i6lflt/local_llama_run_pod_vscode_setup/
Normal-Diver7342
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6lflt
false
null
t3_1i6lflt
/r/LocalLLaMA/comments/1i6lflt/local_llama_run_pod_vscode_setup/
false
false
self
1
null
Few-shot examples with structured output
1
So, I’ve been working on a problem where I have to ingest few-shot examples with structured output. I am currently working with openAI models and they don’t have a direct way around this They do have a work-around with tool/function calling but local models generally don’t support that the easy way I will soon be working with a fine-tuned version of Llama3.3 and want to know in advance of strategies that I can use to add few shot examples with structured outputs I can few shot examples in the system messages and then have user-assistant messages to mimic the exact format that I want but again It does have it’s own downsides (it might not always respond in the structured output format specified) So, what can be done in this case? Any help would be appreciated Thanks ;)
2025-01-21T15:51:23
https://www.reddit.com/r/LocalLLaMA/comments/1i6lmxp/fewshot_examples_with_structured_output/
wtf-beech
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6lmxp
false
null
t3_1i6lmxp
/r/LocalLLaMA/comments/1i6lmxp/fewshot_examples_with_structured_output/
false
false
self
1
null
From llama2 --> DeepSeek R1 things have gone a long way in a 1 year
447
I was blown away by llama2 70b when it came out. I felt so empowered having so much knowledge spun up locally on my M3 Max. Just over a year, and DeepSeek R1 makes Llama 2 seem like a little child. It's crazy how good the outputs are, and how fast it spits out tokens in just 40GB. Can't imagine where things will be in another year.
2025-01-21T15:58:08
https://www.reddit.com/r/LocalLLaMA/comments/1i6lsgo/from_llama2_deepseek_r1_things_have_gone_a_long/
Vegetable_Sun_9225
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6lsgo
false
null
t3_1i6lsgo
/r/LocalLLaMA/comments/1i6lsgo/from_llama2_deepseek_r1_things_have_gone_a_long/
false
false
self
447
null
What is the prompt format for text completion in DeepSeek 3.0?
1
I’ve been searching for a while, and it seems they only support Chat Completion now, not Text Completion. Can someone confirm this? What can I do if I really want to use Text Completion? For example, for Llama, the formatting looks like this is: <|begin_of_text|><|start_header_id|>system<|end_header_id|> You are a helpful AI assistant for travel tips and recommendations<|eot_id|><|start_header_id|>user<|end_header_id|> What can you help me with?<|eot_id|><|start_header_id|>assistant<|end_header_id|> What would it be for DeepSeek? They must have trained it with a specific format, after all. Thanks
2025-01-21T16:06:12
https://www.reddit.com/r/LocalLLaMA/comments/1i6lzq8/what_is_the_prompt_format_for_text_completion_in/
houmie
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6lzq8
false
null
t3_1i6lzq8
/r/LocalLLaMA/comments/1i6lzq8/what_is_the_prompt_format_for_text_completion_in/
false
false
self
1
null
deepseek-r1:14b strawberry test
1
[removed]
2025-01-21T16:10:30
https://www.reddit.com/r/LocalLLaMA/comments/1i6m3at/deepseekr114b_strawberry_test/
Derkugelscheiber
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6m3at
false
null
t3_1i6m3at
/r/LocalLLaMA/comments/1i6m3at/deepseekr114b_strawberry_test/
false
false
self
1
null
ollama vs huggingface versions of deepseek r1 distill qwen 1.5B
0
just asking if anyone has seen deepseek r1 distill qwen 1.5b . on ollama it is 1.9GB and on huggingface 3.8GB something anyone know what is the difference. On ollama those models are not from deepseek officially.
2025-01-21T16:12:44
https://www.reddit.com/r/LocalLLaMA/comments/1i6m535/ollama_vs_huggingface_versions_of_deepseek_r1/
Professional_Helper_
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6m535
false
null
t3_1i6m535
/r/LocalLLaMA/comments/1i6m535/ollama_vs_huggingface_versions_of_deepseek_r1/
false
false
self
0
null
Deploy any LLM on Huggingface at 3-10x Speed
132
2025-01-21T16:30:20
https://i.redd.it/8dsnudtrhdee1.png
avianio
i.redd.it
1970-01-01T00:00:00
0
{}
1i6mjxv
false
null
t3_1i6mjxv
/r/LocalLLaMA/comments/1i6mjxv/deploy_any_llm_on_huggingface_at_310x_speed/
false
false
https://b.thumbs.redditm…7zOszujL1wto.jpg
132
{'enabled': True, 'images': [{'id': 'Gu7rfVyqHNAWFcqxVOuH-zikG0-GURwzGMdsdtu74vQ', 'resolutions': [{'height': 51, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=108&crop=smart&auto=webp&s=d45da0f68ab3e8d28c069755dfbdb07505119184', 'width': 108}, {'height': 103, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=216&crop=smart&auto=webp&s=b9a948b57dbf79e327551c9a6ad22bca4e414683', 'width': 216}, {'height': 153, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=320&crop=smart&auto=webp&s=5fc59fd06d337179ea518a2d0f70f87efbfba76a', 'width': 320}, {'height': 307, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=640&crop=smart&auto=webp&s=4bf725a99b4877d47c42a9a0a7d813a8339eb266', 'width': 640}, {'height': 461, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=960&crop=smart&auto=webp&s=eac46c0e6c7d54177a541d058472c241cf2145c4', 'width': 960}, {'height': 519, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?width=1080&crop=smart&auto=webp&s=ccc3deb39798cccabb4b9523a580a31807c16c09', 'width': 1080}], 'source': {'height': 859, 'url': 'https://preview.redd.it/8dsnudtrhdee1.png?auto=webp&s=22f2daade94c08af2f4adf32769ebb6e69267d3d', 'width': 1785}, 'variants': {}}]}
Hosting Local LLM for large User Base
2
If an organisation wants to host a local llm (eg 32B) for lots of concurrent users what kind of infrastructure would be required? I know the general VRAM requirements for running LLMs but I’m unsure what happens when it processes lots of concurrent queries for inference… Could thus be done on something like a wrx80 with a couple A100s?
2025-01-21T16:39:48
https://www.reddit.com/r/LocalLLaMA/comments/1i6ms70/hosting_local_llm_for_large_user_base/
NewBronzeAge
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6ms70
false
null
t3_1i6ms70
/r/LocalLLaMA/comments/1i6ms70/hosting_local_llm_for_large_user_base/
false
false
self
2
null
Best TTS and STT models that support websocket streaming without too much overhead
4
As the title says, im looking for a good production set of TTS and STT models that support real time websocket streaming. This is going to be initially a pipeline for speech to speech but my team will eventually latently bridge these opensource into a soft end-end pipeline with deepseek in the middle.
2025-01-21T16:50:34
https://www.reddit.com/r/LocalLLaMA/comments/1i6n1i6/best_tts_and_stt_models_that_support_websocket/
Masony817
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6n1i6
false
null
t3_1i6n1i6
/r/LocalLLaMA/comments/1i6n1i6/best_tts_and_stt_models_that_support_websocket/
false
false
self
4
null
Math-Verify: Fixing broken Math Benchmarks
2
[https://x.com/HKydlicek/status/1881734376696041659](https://x.com/HKydlicek/status/1881734376696041659)
2025-01-21T16:53:33
https://www.reddit.com/r/LocalLLaMA/comments/1i6n3zg/mathverify_fixing_broken_math_benchmarks/
Other_Housing8453
self.LocalLLaMA
1970-01-01T00:00:00
0
{}
1i6n3zg
false
null
t3_1i6n3zg
/r/LocalLLaMA/comments/1i6n3zg/mathverify_fixing_broken_math_benchmarks/
false
false
self
2
{'enabled': False, 'images': [{'id': '43RfpbVsl0DVwVG5Af-pu2nT8e4SQJN79k8EKYfcj7s', 'resolutions': [{'height': 63, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?width=108&crop=smart&auto=webp&s=e8d7b5860425aa93a1be3a7610caca2b9f0dd20c', 'width': 108}, {'height': 127, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?width=216&crop=smart&auto=webp&s=8afa6d8c99bc750b4d7efac4a47b238e5d9cc1c7', 'width': 216}, {'height': 188, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?width=320&crop=smart&auto=webp&s=d298bd79ff24dd32503620195ec9d9fecda49b94', 'width': 320}, {'height': 376, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?width=640&crop=smart&auto=webp&s=87fd10f3ba87e77fc87497790e63adccd6c94cc8', 'width': 640}, {'height': 565, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?width=960&crop=smart&auto=webp&s=1cade6bca37aea53232aa23832c8f3c19978c224', 'width': 960}], 'source': {'height': 583, 'url': 'https://external-preview.redd.it/uarGAMlkhC3k7IfFqjUWJOarnl9bmeZh-H3tbGvUyZk.jpg?auto=webp&s=339497a6d1ec94bade34b008dbfd5e8151cac731', 'width': 990}, 'variants': {}}]}