Maziyar PANAHI
@maziyarpanahi.bsky.social
📤 315
📥 52
📝 92
AI x Healthcare | ❤️
#opensource
| e/acc 🇫🇷 Follow me on X:
https://x.com/MaziyarPanahi
pinned post!
🚀 Big news in healthcare AI! I'm thrilled to announce the launch of OpenMed on
@hf.co
, releasing 380+ state-of-the-art medical NER models for free under Apache 2.0. And this is just the beginning! 🧵
huggingface.co/blog/Maziyar...
loading . . .
Unlocking Healthcare AI: I'm Releasing State-of-the-Art Medical Models for Free. Forever.
A Blog post by Maziyar Panahi on Hugging Face
https://huggingface.co/blog/MaziyarPanahi/open-health-ai
6 months ago
1
0
0
Introducing 90+ open-source, state‑of‑the‑art biomedical and clinical zero‑shot NER models on
@hf.co
by OpenMed Apache‑2.0 licensed and ready to use Built on GLiNER and covering 12+ biomedical datasets 🧵 (1/6)
4 months ago
1
0
0
welcome GPT-5-Codex
4 months ago
0
0
0
Unlocking Healthcare AI: I'm Releasing State-of-the-Art Medical Models for Free. Forever.
huggingface.co/blog/Maziyar...
loading . . .
Unlocking Healthcare AI: I'm Releasing State-of-the-Art Medical Models for Free. Forever.
A Blog post by Maziyar Panahi on Hugging Face
https://huggingface.co/blog/MaziyarPanahi/open-health-ai
6 months ago
0
1
0
🚀 Big news in healthcare AI! I'm thrilled to announce the launch of OpenMed on
@hf.co
, releasing 380+ state-of-the-art medical NER models for free under Apache 2.0. And this is just the beginning! 🧵
huggingface.co/blog/Maziyar...
loading . . .
Unlocking Healthcare AI: I'm Releasing State-of-the-Art Medical Models for Free. Forever.
A Blog post by Maziyar Panahi on Hugging Face
https://huggingface.co/blog/MaziyarPanahi/open-health-ai
6 months ago
1
0
0
After 14 years, I'm canceling my Apple Developer membership. I've always believed Apple should pay developers to build apps, not charge them. iPhone is useless without developers' work. Stop taking money from developers; they already ensure your overpriced devices sell!
8 months ago
1
1
0
RL in LLM training is like adding the right spices to a dish - it just makes everything better!
medium.com/%40ardyadipt...
loading . . .
Teaching AI to Understand Us: Reinforcement Learning with Human Feedback (RLHF)
Have you ever experienced where LLM provides you with the wrong results or hallucination? and you wish to correct them so that in the…
https://medium.com/%40ardyadipta/teaching-ai-to-understand-us-reinforcement-learning-with-human-feedback-rlhf-5fc5ae16da40
12 months ago
0
5
1
RLHF is changing the game, making AI more human-like by learning from our feedback, but it's not all easy - getting good, unbiased feedback is tough.
www.lebigdata.fr/tout-savoir-...
loading . . .
Tout savoir sur le RLHF, ou Reinforcement Learning from Human Feedback
Percez le secret du RLHF, la technique qui permet aux modèles IA d'apprendre de leurs erreurs et d'ajuster leurs réponses.
https://www.lebigdata.fr/tout-savoir-sur-le-rlhf-ou-reinforcement-learning-from-human-feedback
12 months ago
0
0
0
Check out Reward-Robust RLHF - it's tackling reward hacking and making LLMs more reliable by focusing on both performance and stability.
medium.com/%40TheDataSc...
loading . . .
DeepSeek-R1: Revolutionizing Reasoning in Large Language Models with Reinforcement Learning
In the rapidly evolving field of artificial intelligence, DeepSeek-AI has introduced a groundbreaking approach to enhancing the reasoning…
https://medium.com/%40TheDataScience-ProF/deepseek-r1-revolutionizing-reasoning-in-large-language-models-with-reinforcement-learning-1220f0f7fb43
12 months ago
0
0
0
New DeepSeek-R1 method boosts LLM reasoning with a cool multi-stage training setup, making AI smarter at problem-solving.
medium.com/%40danushidk...
loading . . .
DeepSeek-R1: Incentivizing Reasoning Capability in Large Language Models via Reinforcement Learning…
With Large Language Models (LLMs) opening the door to artificial general intelligence (AGI), the area of artificial intelligence (AI) has…
https://medium.com/%40danushidk507/deepseek-r1-incentivizing-reasoning-capability-in-large-language-models-via-reinforcement-learning-9515a28a23ad
12 months ago
0
0
0
Now you can use DeepSeek R1 on Azure AI Foundry and GitHub, making top-notch AI tech more accessible to everyone.
12 months ago
1
0
0
Sam Altman admits OpenAI was wrong about open-source AI. Guess they're playing catch-up now!
www.businessinsider.com/sam-altman-o...
loading . . .
Sam Altman says OpenAI will embrace two new AI approaches, one from DeepSeek and another from Meta
During a Reddit AMA on Friday, Altman said OpenAI has "been on the wrong side of history" when it comes to keeping model weights confidential.
https://www.businessinsider.com/sam-altman-openai-ai-approaches-deepseek-meta-open-source-2025-1
12 months ago
0
0
0
reposted by
Maziyar PANAHI
rat king 🐀 👑
12 months ago
weird thing of a bunch of spammers in my mentions that are all obviously llm replies all accounts that have existed for less than 24 hours, only replies and no original posts, and oddly fixated on social issues
23
434
54
Building reasoning models is no easy feat! 🚀 DeepSeek-R1’s journey highlights key challenges with PRM (Process Reward Models) and MCTS (Monte Carlo Tree Search). From annotation hurdles to scaling limits, the path to scalable AI reasoning is full of learnings.
#AI
#ReinforcementLearning
#rl
12 months ago
1
0
0
Thank you! ☺️ love a strong dollar! 😆
about 1 year ago
0
0
0
The next season of Selling Sunset is going to be interesting!
add a skeleton here at some point
about 1 year ago
0
0
0
“control the text generation process itself by directly modifying the probability distribution? That’s where logit processing comes into play.” Read more on
@hf.co
:
huggingface.co/blog/logits-...
loading . . .
Controlling Language Model Generation with NVIDIA's LogitsProcessorZoo
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
https://huggingface.co/blog/logits-processor-zoo
about 1 year ago
0
2
0
reposted by
Maziyar PANAHI
Anton
about 1 year ago
Introducing 📐FineMath: the best open math pre-training dataset with 50B+ tokens! Math remains challenging for LLMs and by training on FineMath we see considerable gains over other math datasets, especially on GSM8K and MATH. 🤗
huggingface.co/datasets/Hug...
Here’s a breakdown 🧵
2
46
16
o1 still not compatible with web search is a big problem!
about 1 year ago
0
0
0
What the hell happened to Archer on Netflix_France? Out of 13 seasons we now only have 2! We are not getting anything new, now we are losing old ones we liked. It's time to cancel until there is something to watch.
about 1 year ago
0
0
0
Can we talk about why LinkedIn Premium is so expensive? They charge $69 a month just to help you grow your network! What’s the real added value that justifies such a high price? (I remember when Premium was just $10 a month!)
about 1 year ago
0
0
0
Check out the new Open LLM Leaderboard by
@hf.co
❤️
huggingface.co/spaces/open-...
loading . . .
Open LLM Leaderboard - a Hugging Face Space by open-llm-leaderboard
Track, rank and evaluate open LLMs and chatbots
https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=qwen2&pinned=LGAI-EXAONE%2FEXAONE-3.5-7.8B-Instruct&official=true
about 1 year ago
0
0
0
Jay Taj deserved to win! 🏆 “where is you from?”
about 1 year ago
0
2
0
It's not the best, but still nice to be able to use VSCode/Cursor over SSH. I rarely run any code on my macbook anymore!
about 1 year ago
0
1
0
temperature 0 vs. 0.001 😆
about 1 year ago
0
0
0
It's almost 2025, OpenAI has a $200 subscription plan, and you still CANNOT star (favorite) a conversation like Claude!
about 1 year ago
0
2
0
reposted by
Maziyar PANAHI
Nathan Cooper
about 1 year ago
As R&D staff @
answer.ai
, I work a lot on boosting productivity with AI. A common theme that always comes up is the combination of human+AI. This combination proved to be powerful in our new project ShellSage, which is an AI terminal buddy that learns and teaches with you. A 🧵
7
72
23
That’s all I’m gonna say!
about 1 year ago
0
0
0
New GGUF model drop alert!
huggingface.co/MaziyarPanah...
loading . . .
MaziyarPanahi/Sailor2-20B-Chat-GGUF · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
https://huggingface.co/MaziyarPanahi/Sailor2-20B-Chat-GGUF
about 1 year ago
1
1
0
reposted by
Maziyar PANAHI
Michael Tschannen
about 1 year ago
Have you ever wondered how to train an autoregressive generative transformer on text and raw pixels, without a pretrained visual tokenizer (e.g. VQ-VAE)? We have been pondering this during summer and developed a new model: JetFormer 🌊🤖
arxiv.org/abs/2411.19722
A thread 👇 1/
4
155
43
reposted by
Maziyar PANAHI
vb
about 1 year ago
Introducing Indic-Parler TTS - Trained on 10K hours of data, 938M params, supports 20 Indic languages, emotional synthesis, apache 2.0 licensed! 🔥 w/ fully customisable speech and voice personas! Try it out directly below or use the model weights as you want! 🇮🇳/acc
4
35
3
reposted by
Maziyar PANAHI
Tom Aarsen
about 1 year ago
Check out the most liked and downloaded models from 2022 to 2024 on Hugging Face:
huggingface.co/spaces/huggi...
Tomorrow we'll have another visualization!
2
15
3
I say one word, you complete the rest: Reflection …
about 1 year ago
1
0
0
Just finished season 2 of Diplomat on Netflix. What an ending! 😂 ( Kate, I did something bad!)
about 1 year ago
0
1
0
reposted by
Maziyar PANAHI
Loubna Ben Allal
about 1 year ago
We hit 1K ⭐ on our SmolLM repo—thank you! 🎉 New updates: • SmolLM2 nanotron checkpoints (with optimizer states) for easier continual pre-training • Local inference demos (MLC, Transformers.js, MLX, llama.cpp) • SmolVLM: Vision-language model built on SmolLM2
github.com/huggingface/...
0
18
1
Who’s playing the new Star Wars Outlaws on PS5?
about 1 year ago
0
0
0
There’s no room for toxicity or harsh comments. Honestly, if it were me, I might not have removed it, let alone apologized! So, let’s be grateful that Daniel is a much better person than most of us! Let's discuss what should be the rules when platforms like BlueSky or Mastodon have open data?
add a skeleton here at some point
about 1 year ago
0
3
0
And here I thought Anthropic was adding response styles as features! It turns out it’s just a way to switch to shorter responses to save tokens!
about 1 year ago
0
4
0
reposted by
Maziyar PANAHI
Simon Willison
about 1 year ago
Some notes on the new AWS S3 conditional write feature where you can use an if-match header to avoid over-writing an object if it's been changed since you last read it
simonwillison.net/2024/Nov/26/...
loading . . .
Amazon S3 adds new functionality for conditional writes
> Amazon S3 can now perform conditional writes that evaluate if an object is unmodified before updating it. This helps you coordinate simultaneous writes to the same object and prevents …
https://simonwillison.net/2024/Nov/26/s3-conditional-writes/
1
73
6
Which one do you prefer and why? 😀 1. L40S (48G) 2. RTX A6000 (48G) To serve LLMs in production.
about 1 year ago
1
0
0
Are you kidding me!? Back to JSON I guess! 😅 "Does Prompt Formatting Have Any Impact on LLM Performance?" -
arxiv.org/pdf/2411.10541
about 1 year ago
2
5
2
Challenge accepted! total_num_tokens: 2_438_197_708
about 1 year ago
1
0
0
reposted by
Maziyar PANAHI
Simon Willison
about 1 year ago
I like this new analogy for working with LLMs by
@emollick.bsky.social
"treat AI like an infinitely patient new coworker who forgets everything you tell them each new conversation, one that comes highly recommended but whose actual abilities are not that clear"
www.oneusefulthing.org/p/getting-st...
loading . . .
Getting started with AI: Good enough prompting
Don't make this hard
https://www.oneusefulthing.org/p/getting-started-with-ai-good-enough
6
184
25
A team behind SmolLM2 model at
@huggingface.bsky.social
just released everything! A true open-source AI: - Pre-training code - Evaluation suite - Synthetic data generation - Post-training scripts with TRL - On-device tools for summarization, rewriting & agents All with Apache 2.0 licensed! 🔥
about 1 year ago
1
38
10
The safety alignments on Claude is so stupid that sometimes I just want to cancel my subscription! When it comes to questions Claude refuses to answer then OpenAI is the king!
about 1 year ago
0
0
0
“In a stunning misstep, OpenAI engineers accidentally erased critical evidence gathered by The New York Times and other major newspapers in their lawsuit over AI training data, according to a court filing Wednesday.” 😂
about 1 year ago
0
3
0
Having fun with batch sizes, from bottom to top: • Gradient steps: 16 • Micro batch size: 4 • Gradient steps: 8 • Micro batch size: 3 • Gradient steps: 16 • Micro batch size: 3
about 1 year ago
0
0
0
Casual day in Paris! Snow & LLMs! ❄️
about 1 year ago
0
1
0
reposted by
Maziyar PANAHI
Michael Bang Petersen
about 1 year ago
A mapping of how Bluesky is becoming the new Scientific Twitter
mikeyoungacademy.dk/bluesky-is-e...
loading . . .
Bluesky is emerging as the new platform for science - Mike Young Academy
Scientific Twitter is about to find its true successor. And it is not X. This, our latest release, shows that the Bluesky network of scientists is growing — and growing.
https://mikeyoungacademy.dk/bluesky-is-emerging-as-the-new-platform-for-science/
37
2350
774
Don't forget the CALME-3 models!🇫🇷🥳 My 3B LLM collection on
@huggingface.bsky.social
: • 3 new instruction-tuned models • 3 French language models (Project Baguette) • 6 specialized French legal models (LoiLlama & LoiQwen) • 54M-token French legal synthetic dataset
about 1 year ago
1
2
0
"A Comprehensive Survey of Small Language Models in the Era of Large Language Models: Techniques, Enhancements, Applications, Collaboration with LLMs, and Trustworthiness" LLMs face size and privacy limits. SLMs are alternatives for specialized tasks and edge use.
arxiv.org/pdf/2411.03350
loading . . .
https://arxiv.org/pdf/2411.03350
about 1 year ago
0
1
0
Load more
feeds!
log in