Tag: Qwen
-
Simon Willison’s Weblog: TextSynth Server
Source URL: https://simonwillison.net/2024/Nov/21/textsynth-server/ Source: Simon Willison’s Weblog Title: TextSynth Server Feedly Summary: TextSynth Server I’d missed this: Fabrice Bellard (yes, that Fabrice Bellard) has a project called TextSynth Server which he describes like this: ts_server is a web server proposing a REST API to large language models. They can be used for example for text…
-
Hacker News: Qwen2.5 Turbo extends context length to 1M tokens
Source URL: http://qwenlm.github.io/blog/qwen2.5-turbo/ Source: Hacker News Title: Qwen2.5 Turbo extends context length to 1M tokens Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the introduction of Qwen2.5-Turbo, a large language model (LLM) that significantly enhances processing capabilities, particularly with longer contexts, which are critical for many applications involving AI-driven natural language…
-
Simon Willison’s Weblog: Qwen: Extending the Context Length to 1M Tokens
Source URL: https://simonwillison.net/2024/Nov/18/qwen-turbo/#atom-everything Source: Simon Willison’s Weblog Title: Qwen: Extending the Context Length to 1M Tokens Feedly Summary: Qwen: Extending the Context Length to 1M Tokens The new Qwen2.5-Turbo boasts a million token context window (up from 128,000 for Qwen 2.5) and faster performance: Using sparse attention mechanisms, we successfully reduced the time to first…
-
Simon Willison’s Weblog: Quoting Jack Clark
Source URL: https://simonwillison.net/2024/Nov/18/jack-clark/ Source: Simon Willison’s Weblog Title: Quoting Jack Clark Feedly Summary: The main innovation here is just using more data. Specifically, Qwen2.5 Coder is a continuation of an earlier Qwen 2.5 model. The original Qwen 2.5 model was trained on 18 trillion tokens spread across a variety of languages and tasks (e.g, writing,…
-
Simon Willison’s Weblog: NuExtract 1.5
Source URL: https://simonwillison.net/2024/Nov/16/nuextract-15/#atom-everything Source: Simon Willison’s Weblog Title: NuExtract 1.5 Feedly Summary: NuExtract 1.5 Structured extraction – where an LLM helps turn unstructured text (or image content) into structured data – remains one of the most directly useful applications of LLMs. NuExtract is a family of small models directly trained for this purpose, and released…
-
Hacker News: Something weird is happening with LLMs and Chess
Source URL: https://dynomight.net/chess/ Source: Hacker News Title: Something weird is happening with LLMs and Chess Feedly Summary: Comments AI Summary and Description: Yes Summary: This text discusses an exploration of how various large language models (LLMs) perform at playing chess, ultimately revealing significant differences in performance across models. Despite enthusiasm about LLMs’ capabilities, the results…
-
Simon Willison’s Weblog: Qwen2.5-Coder-32B is an LLM that can code well that runs on my Mac
Source URL: https://simonwillison.net/2024/Nov/12/qwen25-coder/ Source: Simon Willison’s Weblog Title: Qwen2.5-Coder-32B is an LLM that can code well that runs on my Mac Feedly Summary: There’s a whole lot of buzz around the new Qwen2.5-Coder Series of open source (Apache 2.0 licensed) LLM releases from Alibaba’s Qwen research team. On first impression it looks like the buzz…