- 00:11 β Introduction of Qwen 3.5-9b-Sushi-Coder by Big Tuna AI Labs. @tunahorse21
- 00:17 β Pi prompt templates can now loop and rotate between different models. @nicopreme
- 00:11 β Development of a clinic concierge demo using GPT technology. @OpenAIDevs
- 00:02 β Discussion on the capabilities of GLM-5.1 model. @aicodeking
- 00:19 β Announcement of improvements in Hermes Agent integration with Hugging Face. @NousResearch
- 00:47 β Overview of a coding model suitable for low RAM devices. @0xSero
- 02:13 β Warning about GitHub’s use of user code and data for AI training. @kaivara
- 02:05 β Positive feedback on Qwen 3.5 27B model performance. @TheAhmadOsman
- 00:19 β TurboQuant CUDA for llama.cpp: 3.5x KV cache compression that beats q8_0 quality. @spiritbuun
- 00:27 β CLI agents using CLI browsers, potentially the future of AI browser use. @shawn_pana
AI & Tech Developments - Mar 28
π± Source Tweets
breaking news big tuna ai labs presents Qwen3.5-9b-Sushi-Coder sft on opus reasoning traces and codeforces with unsloth then RLed with hermes agent using Atropos for coding RL im not raising !
β @tunahorse21
Pi prompt templates can now loop and rotate between different models on each iteration. In this example it runs each loop with a different model as a subagent based on a fork of the current chat. pi install npm:pi-prompt-template-model https://github.com/nicobailon/pi-prompt-template-modelβ¦
β @nicopreme
Build voice agents that do real work. We built a clinic concierge demo for a Singapore health clinic with gpt-realtime-1.5. It speaks naturally with patients, collects the right details, and books appointments in real time.
β @OpenAIDevs
GLM-5.1 by @Zai_org is one of the best agentic models out there. I've been testing it early and it is genuienly impressive. Way better at instruction following, long running tasks than previous generation. full review here: https://youtu.be/UxGieu7PaPg
β @aicodeking
We have integrated @huggingface as a first-class inference provider in Hermes Agent. When you select Hugging Face in the model picker it now shows 28 curated models organized by use case, with a custom option for the 100+ other models they serve.
β @NousResearch
For everyone with 16 or less GB ram that want to run a coding model locally. This is an autocomplete code model that Iβve noticed does pretty well. Not as good as Cursor tab but good enough
β @0xSero
GIVE ME YOUR ATTENTION FOR LIKE 5 SECONDS GITHUB IS GONNA START USING *YOUR* CODE AND DATA TO TRAIN AI USING COPILOT!!!!!!!!!!! DISABLE IT AT http://github.com/settings/copilot/featuresβ¦
β @_kaivara_
Qwen 3.5 27B (Dense) with Hermes Agent is REALLY GOOD
β @TheAhmadOsman
Introduction of Qwen 3.5-9b-Sushi-Coder by Big Tuna AI Labs.
β @tunahorse21
Pi prompt templates can now loop and rotate between different models.
β @nicopreme
Development of a clinic concierge demo using GPT technology.
β @OpenAIDevs
Discussion on the capabilities of GLM-5.1 model.
β @aicodeking
Announcement of improvements in Hermes Agent integration with Hugging Face.
β @NousResearch
Overview of a coding model suitable for low RAM devices.
β @0xSero
Warning about GitHub's use of user code and data for AI training.
β @_kaivara_
Positive feedback on Qwen 3.5 27B model performance.
β @TheAhmadOsman
TurboQuant CUDA for llama.cpp: 3.5x KV cache compression that beats q8_0 quality.
β @spiritbuun
CLI agents using CLI browsers, potentially the future of AI browser use.
β @shawn_pana