@anon
sign up
@anon
sign up
pull down to refresh
SlowLlama: Finetune llama2-70B and codellama on MacBook Air without quantization
github.com/okuvshynov/slowllama
10 sats
\
1 comment
\
@hn
6 Oct 2023
tech
related
OpenAI offers free GPT-4o Mini fine-tuning to counter Meta’s Llama 3.1 release
venturebeat.com/ai/ai-arms-race-escalates-openai-offers-free-gpt-4o-mini-fine-tuning-to-counter-metas-llama-3-1-release/
21 sats
\
0 comments
\
@ch0k1
25 Jul 2024
news
Show HN: 80% faster, 50% less memory, 0% loss of accuracy Llama finetuning
github.com/unslothai/unsloth
21 sats
\
2 comments
\
@hn
2 Dec 2023
tech
LLaMA-Factory: Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
github.com/hiyouga/LLaMA-Factory
157 sats
\
0 comments
\
@carter
19 Sep
AI
Orchard - Lightning, Cashu, Tether, Llama
orchard.space/
427 sats
\
5 comments
\
@Scoresby
24 Jun
lightning
cocktailpeanut/dalai: The simplest way to run LLaMA on your local machine
github.com/cocktailpeanut/dalai
247 sats
\
0 comments
\
@random_
24 Mar 2023
bitcoin
Qwen2.5-Coder-32B is an LLM that can code well that runs on my Mac
simonwillison.net/2024/Nov/12/qwen25-coder/
19 sats
\
0 comments
\
@Rsync25
13 Nov 2024
tech
How to Run Llama 3.1 405B on Home Devices? Build AI Cluster!
b4rtaz.medium.com/how-to-run-llama-3-405b-on-home-devices-build-ai-cluster-ad0d5ad3473b
116 sats
\
3 comments
\
@Rsync25
29 Jul 2024
alter_native
Minimal implementation of Mamba, the new LLM architecture, in 1 file of PyTorch
github.com/johnma2006/mamba-minimal
15 sats
\
1 comment
\
@hn
20 Dec 2023
tech
Hardware Acceleration of LLMs: A comprehensive survey and comparison
arxiv.org/abs/2409.03384
21 sats
\
0 comments
\
@hn
7 Sep 2024
tech
Apple collaborates with NVIDIA to research faster LLM performance - 9to5Mac
9to5mac.com/2024/12/18/apple-collaborates-with-nvidia-to-research-faster-llm-performance/
14 sats
\
1 comment
\
@Rsync25
19 Dec 2024
tech
Accelerating the Performance of Rosetta in Linux VMs on Apple Silicon
developer.apple.com/documentation/virtualization/accelerating_the_performance_of_rosetta
10 sats
\
0 comments
\
@hn
7 Nov 2024
tech
Compiling LLMs into a MegaKernel: A path to low-latency inference
zhihaojia.medium.com/compiling-llms-into-a-megakernel-a-path-to-low-latency-inference-cf7840913c17
10 sats
\
0 comments
\
@hn
19 Jun
tech
The Best Way of Running GPT-OSS Locally - KDnuggets
www.kdnuggets.com/the-best-way-of-running-gpt-oss-locally
118 sats
\
0 comments
\
@optimism
25 Aug
AI
Meet PowerInfer: A Fast LLM on a Single Consumer-Grade GPU
www.marktechpost.com/2023/12/23/meet-powerinfer-a-fast-large-language-model-llm-on-a-single-consumer-grade-gpu-that-speeds-up-machine-learning-model-inference-by-11-times/
10 sats
\
2 comments
\
@ch0k1
24 Dec 2023
AI
Run LLMs on my own Mac fast and efficient Only 2 MBs
www.secondstate.io/articles/fast-llm-inference/
13 sats
\
1 comment
\
@hn
13 Nov 2023
tech
LND v0.16.1-beta: Bug Fixes and Optimizations
github.com/lightningnetwork/lnd/blob/master/docs/release-notes/release-notes-0.16.1.md
6739 sats
\
0 comments
\
@kepford
25 Apr 2023
bitcoin
New Ultra Fast Lossless Audio Codec (HALAC)
hydrogenaud.io/index.php/topic,125248.0.html?PHPSESSID=su12fvq4ar7cordpr8p1uoa7p0
33 sats
\
1 comment
\
@hn
2 Jan 2024
tech
How Is LLaMa.cpp Possible?
finbarr.ca/how-is-llama-cpp-possible/
16 sats
\
2 comments
\
@hn
15 Aug 2023
tech
Announcing LND 0.16 beta: Smarter, Leaner, Faster
lightning.engineering/posts/2023-03-29-lnd-0.16-launch/
20 sats
\
0 comments
\
@AR0w
27 Apr 2023
bitcoin
Bringing K/V context quantisation to Ollama
smcleod.net/2024/12/bringing-k/v-context-quantisation-to-ollama/
11 sats
\
0 comments
\
@hn
5 Dec 2024
tech
1-Bit LLM: The Most Efficient LLM Possible?
www.youtube.com/watch?v=7hMoz9q4zv0
533 sats
\
1 comment
\
@carter
24 Jun
AI
more