Un conjunto de modelos realmente buenos para el edge se lanzó esta semana - Jan lanza un modelo de 4B optimizado para ejecutar tareas agentivas localmente - supera a Perplexity Pro en varias evaluaciones - LiquidAI lanza modelos multimodales para texto+imagen con versiones de 450M y 1.6B optimizadas para baja latencia
👋 Jan
👋 Jan12 ago, 15:55
Introducing Jan-v1: 4B model for web search, an open-source alternative to Perplexity Pro. In our evals, Jan v1 delivers 91% SimpleQA accuracy, slightly outperforming Perplexity Pro while running fully locally. Use cases: - Web search - Deep Research Built on the new version of Qwen's Qwen3-4B-Thinking (up to 256k context length), fine-tuned for reasoning and tool use in Jan. You can run the model in Jan, llama.cpp, or vLLM. To enable search in Jan, go to Settings → Experimental Features → On, then Settings → MCP Servers → enable a search-related MCP such as Serper. Use the model: - Jan-v1-4B: - Jan-v1-4B-GGUF: Credit to the @Alibaba_Qwen team for Qwen3 4B Thinking & @ggerganov for llama.cpp.
10,85K