Erik
@erikkaum.bsky.social
π€ 830
π₯ 133
π 38
SWE
@hf.co
We have Nvidia B200s ready to go for you in Hugging Face Inference Endpoints π₯ I tried them out myself and the performance is amazing. On top of that we just got a fresh batch of H100s as well. At $4.5/hour it's a clear winner in terms of price/perf compared to the A100.
about 2 months ago
0
6
1
We just refreshed π our analytics in
@hf.co
endpoints. More info below!
loading . . .
8 months ago
1
8
3
Morning workout at the
@hf.co
Paris office is imo one of the best perks.
9 months ago
0
3
0
Gemma 3 is live π₯ You can deploy it from endpoints directly with an optimally selected hardware and configurations. Give it a try π
loading . . .
9 months ago
1
7
3
Apparently, mom is a better engineer than what I am.
11 months ago
0
4
0
today as part of a course, I implemented a program that takes a bit stream like so: 10001001110111101000100111111011 and decodes the intel 8088 assembly from it like: mov si, bx mov bx, di only works on the mov instruction, register to register. code:
github.com/ErikKaum/bit...
loading . . .
https://github.com/ErikKaum/bitbubble
https://t.co/fcNDebb0fk
11 months ago
0
1
0
Ambition is a paradox. You should always aim higher, but that easily becomes a state where you're never satisfied. Just reached 10k MRR. Now there's the next goal of 20k. Sharif has a good talk on this: emotional runway. How do you deal with this paradox? video:
www.youtube.com/watch?v=zUnQ...
loading . . .
before you give up, give this video a chance.
YouTube video by Founders, Inc.
https://www.youtube.com/watch?v=zUnQYkAiwWU
12 months ago
0
1
0
Qui Gon Jinn sharing some insightful prompting wisdom ππΌ
12 months ago
1
11
3
it's this time of the year π
12 months ago
0
2
0
reposted by
Erik
Xuan Son Nguyen
about 1 year ago
Hugging Face inference endpoints now support CPU deployment for llama.cpp π π Why this is a huge deal? Llama.cpp is well-known for running very well on CPU. If you're running small models like Llama 1B or embedding models, this will definitely save tons of money π° π°
3
23
7
reposted by
Erik
Andi
about 1 year ago
Let's go! We are releasing SmolVLM, a smol 2B VLM built for on-device inference that outperforms all models at similar GPU RAM usage and tokens throughputs. SmolVLM can be fine-tuned on a Google collab and be run on a laptop! Or process millions of documents with a consumer GPU!
4
104
26
Is it just me or does it intuitively align that chat bars are at the bottom of the page and search bars at the top? I've noticed that perplexity positions the question on the top and generates the text below. Is it because they want to position more as a search engine?
about 1 year ago
0
1
0
code boxes with syntax highlighting π
add a skeleton here at some point
about 1 year ago
1
4
0
typical engineer writing copy in plain english i'd say "2 conversions at the same time"
about 1 year ago
0
4
0
lesson: if you care about the performance of something, you gotta run your own benchmarks
add a skeleton here at some point
about 1 year ago
0
10
4
Just wrote some golang for fun. Damn, I had almost forgotten how enjoyable itβs to program in. Just breezing through the code. If I need thousands of threads, itβs just there.
about 1 year ago
0
12
0
A while ago I started experimenting with compiling the Python interpreter to WASM. To build a secure, fast, and lightweight sandbox for code execution β ideal for running LLM-generated Python code. - Send code simply as a POST request - 1-2ms startup times
github.com/ErikKaum/run...
loading . . .
GitHub - ErikKaum/runner: Experimental wasm32-unknown-wasi runtime for Python code execution
Experimental wasm32-unknown-wasi runtime for Python code execution - ErikKaum/runner
https://github.com/ErikKaum/runner
about 1 year ago
0
4
4
There are now /llms.txt files for a few of
@huggingface.bsky.social
docs π₯
huggingface-projects-docs-llms-txt.hf.space/transformers...
loading . . .
http://huggingface-projects-docs-llms-txt.hf.space/transformers/llms.txt
about 1 year ago
0
1
0
@huggingface.bsky.social
off-site π€πΌ
about 1 year ago
0
4
0
"If you're thinking without writing, you only think you're thinking." Same applies imo to coding and why it's so important to open your editor, start tinkering and sketching things out. quote from:
paulgraham.com/writes.html
loading . . .
Writes and Write-Nots
https://paulgraham.com/writes.html
about 1 year ago
0
2
0
you reached the end!!
feeds!
log in