@anon
sign up
@anon
sign up
pull down to refresh
Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity
arxiv.org/abs/2510.01171
147 sats
\
0 comments
\
@carter
16 Oct
AI
related
Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity
arxiv.org/abs/2510.01171
166 sats
\
0 comments
\
@Scoresby
17 Oct
AI
Context Rot: How Increasing Input Tokens Impacts LLM Performance
research.trychroma.com/context-rot
304 sats
\
2 comments
\
@Scoresby
14 Jul
AI
NVIDIA: Transforming LLM Alignment with Efficient Reinforcement Learning
www.marktechpost.com/2024/05/05/nvidia-ai-open-sources-nemo-aligner-transforming-large-language-model-alignment-with-efficient-reinforcement-learning/
20 sats
\
0 comments
\
@ch0k1
7 May 2024
tech
Deep Dive into LLMs like ChatGPT
www.youtube.com/watch?v=7xTGNNLPyMI
98 sats
\
1 comment
\
@kepford
6 May
AI
LLoms - A simple mcp enabled llm cli chat
github.com/gzuuus/lloms
155 sats
\
0 comments
\
@gzuuus_
16 Mar
nostr
Here’s What’s Really Going On Inside An LLM’s Neural Network
116 sats
\
0 comments
\
@0xbitcoiner
22 May 2024
BooksAndArticles
Sampling and structured outputs in LLMs
parthsareen.com/blog.html#sampling.md
157 sats
\
0 comments
\
@carter
23 Sep
AI
Streaming LLM – No limit on context length for your favourite LLM
github.com/mit-han-lab/streaming-llm
10 sats
\
1 comment
\
@hn
2 Oct 2023
tech
Fine-Tuning Increases LLM Vulnerabilities and Risk
arxiv.org/abs/2404.04392
21 sats
\
0 comments
\
@hn
12 Apr 2024
tech
LLM in a Flash: Efficient LLM Inference with Limited Memory
huggingface.co/papers/2312.11514
13 sats
\
1 comment
\
@hn
20 Dec 2023
tech
Are LLMs random?
rnikhil.com/2025/04/26/llm-coin-toss-odd-even
269 sats
\
1 comment
\
@carter
30 Apr
AI
Researchers discover impressive learning capabilities in long-context LLMs
venturebeat.com/ai/deepmind-researchers-discover-impressive-learning-capabilities-in-long-context-llms/
297 sats
\
0 comments
\
@ch0k1
25 Apr 2024
tech
Signal debunks the rumors of LLM integration in the best way possible
317 sats
\
0 comments
\
@StillStackinAfterAllTheseYears
17 Jun
tech
Compiling LLMs into a MegaKernel: A path to low-latency inference
zhihaojia.medium.com/compiling-llms-into-a-megakernel-a-path-to-low-latency-inference-cf7840913c17
10 sats
\
0 comments
\
@hn
19 Jun
tech
Things we learned about LLMs in 2024
simonwillison.net/2024/Dec/31/llms-in-2024/
370 sats
\
0 comments
\
@Rsync25
31 Dec 2024
tech
LLMs generate slop because they avoid surprises by design - Dan Fabulich
danfabulich.medium.com/llms-tell-bad-jokes-because-they-avoid-surprises-7f111aac4f96
343 sats
\
2 comments
\
@Scoresby
19 Aug
AI
Communication Efficient LLM Pre-training with SparseLoCo
arxiv.org/abs/2508.15706
100 sats
\
0 comments
\
@carter
1 Sep
AI
LiveBench - A Challenging, Contamination-Free LLM Benchmark
livebench.ai
161 sats
\
0 comments
\
@supratic
17 Jul
AI
LLM evaluation at scale with the NeurIPS Efficiency Challenge
blog.mozilla.ai/exploring-llm-evaluation-at-scale-with-the-neurips-large-language-model-efficiency-challenge/
110 sats
\
0 comments
\
@localhost
22 Feb 2024
tech
LLaMA-Factory: Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
github.com/hiyouga/LLaMA-Factory
157 sats
\
0 comments
\
@carter
19 Sep
AI
Hardware Acceleration of LLMs: A comprehensive survey and comparison
arxiv.org/abs/2409.03384
21 sats
\
0 comments
\
@hn
7 Sep 2024
tech
more