all solved by qwen 2.5 32B
Anyone using llama 3.2 3b in a flutter app?
Is there really no way you can run 70b models without having a very fast GPU or a lot of ram?
What is the point of these super-tiny LLMs? Can they actually DO anything useful? Here's a short sample interaction with SmoLLM 135m.
Llama 3.2 in production
How to integrate/migrate to loco.rs from Axum app?
What’s the best framework for web servers nowadays?
Would love your input! - Designing MLOps Stack from scratch
[ Removed by Reddit ]
"Generative AI will Require 80% of Engineering Workforce to Upskill Through 2027"
Duration to learn MLOPS
Auto-tuning RAG Models With Katib In Kubeflow
Requesting Feedback on the Feast Kubernetes Operator (the Open Source ML Feature Store)
Llama3 re-write from Pytorch to JAX
Difference between ML Engineering and MLOps
Open Data Lake House with Apache Iceberg and MLOps with Kubeflow
Been married 6 years still a virgin.
Is it just me or are "pure" MLOps roles not that common?
Large Language Model Operations (LLMOps) Specialization
Rust developers what have you done?
Enough already. If I can’t run it in my 3090, I don’t want to hear about it.
Marry a women with a past or a feminist?
Which library to use with Axum for OpenApi?