#languagemodel
Can Socratic reflection improve #AI answers to medical questions?

Adding a critic to a #languageModel pipeline improved performance on two measures of medical question-answering.

The improvement didn't depend on the critic's model.

doi.org/10.48550/arX...

#tech #medicine #edu
January 23, 2026 at 11:58 AM
Forgive me Father, for I have sinned.
January 15, 2026 at 7:39 PM
JMIR Formative Res: Evaluating Spanish Translations of Emergency Department Discharge Instructions by a Large Language Model: Tool Validation and Reliability Study #SpanishTranslations #EmergencyMedicine #HealthcareResearch #LanguageModel #MedicalInterpreting
Evaluating Spanish Translations of Emergency Department Discharge Instructions by a Large Language Model: Tool Validation and Reliability Study
When given a sample of 100 emergency department discharge instructions, Claude Sonnet, a large language model, produced accurate Spanish translations as evaluated by Spanish-speaking physicians and medical interpreters.
dlvr.it
January 12, 2026 at 4:56 PM
Discover the power of AI translation with OpenAI GPT-4! Dive into advanced language understanding, context, and creativity. #LanguageModel #OpenAI #AI
January 10, 2026 at 4:02 PM
Claude Codeについての書籍 "実践Claude Code入門 | 技術評論社" https://gihyo.jp/book/2026/978-4-297-15354-0 #LanguageModel #book
December 17, 2025 at 1:13 PM
This week is another chance to equalize opportunity. To design for inclusion. To make sure no one is left behind by technology.

Welcome to the week. Let’s do today’s work with tomorrow in mind.

#EqualyzAI #NewWeek #MondayMotivation #LanguageModel
December 15, 2025 at 1:24 PM
JMIR Formative Res: Large Language Model Evaluation in Traditional Chinese Medicine for Stroke: Quantitative Benchmarking Study #TraditionalChineseMedicine #TCM #StrokeRecovery #LanguageModel #HealthcareInnovation
Large Language Model Evaluation in Traditional Chinese Medicine for Stroke: Quantitative Benchmarking Study
Background: The application of large language models (LLMs) in medicine is rapidly advancing. However, evaluating LLM capabilities in specialized domains such as traditional Chinese medicine (TCM), which possesses a unique theoretical system and cognitive framework, remains a sizable challenge. Objective: This study aimed to provide an empirical evaluation of different LLM types in the specialized domain of TCM stroke. Methods: The Traditional Chinese Medicine-Stroke Evaluation Dataset (TCM-SED), a 203-question benchmark, was systematically constructed. The dataset includes 3 paradigms (short-answer questions, multiple-choice questions, and essay questions) and covers multiple knowledge dimensions, including diagnosis, pattern differentiation and treatment, herbal formulas, acupuncture, interpretation of classical texts, and patient communication. Gold standard answers were established through a multiexpert cross-validation and consensus process. The TCM-SED was subsequently used to comprehensively test 2 representative LLM models: GPT-4o (a leading international general-purpose model) and DeepSeek-R1 (a large model primarily trained on Chinese corpora). Results: The test results revealed a differentiation in model capabilities across cognitive levels. In objective sections emphasizing precise knowledge recall, DeepSeek-R1 comprehensively outperformed GPT-4o, achieving an accuracy lead of more than 17% in the multiple-choice section (96/137, 70.1% vs 72/137, 52.6%, respectively). Conversely, in the essay section, which tested knowledge integration and complex reasoning, GPT-4o’s performance notably surpassed that of DeepSeek-R1. For instance, in the interpretation of classical texts category, GPT-4o achieved a scoring rate of 90.5% (181/200), far exceeding DeepSeek-R1 (147/200, 73.5%). Conclusions: This empirical study demonstrates that Chinese-centric models have a substantial advantage in static knowledge tasks within the TCM domain, whereas leading general-purpose models exhibit stronger dynamic reasoning and content generation capabilities. The TCM-SED, developed as the benchmark for this study, serves as an effective quantitative tool for evaluating and selecting appropriate LLMs for TCM scenarios. It also offers a valuable data foundation and a new research direction for future model optimization and alignment.
dlvr.it
December 11, 2025 at 4:52 PM
Skąd biorą się dane, na których uczone są modele AI i dlaczego to one często decydują o jakości modeli?

Zapraszamy do przeczytania naszego nowego artykułu o datasetach, transparentności i etyce danych w AI - azurro.pl/skad-biora-s...

#innovation #ArtificialIntelligence #LLM #AI #languagemodel
October 20, 2025 at 11:22 AM
Obfuscating text with local LLMs + CSS highlight API

needed a way after having to blur my address in the Pizza video

Uses Chrome's experimental LanguageModel API to prompt a local v3Nano

Kinda slow, but makes me think of new possibilities when prompting on-device will be fast + free + private
October 17, 2025 at 8:10 PM
Interesting how ChatGPT knows so much about things I know nothing about and is wrong about 70% of the time on topics I'm an expert in. #chatgpt #ai #artificialinteligence #googlegemini #microsoftcopilot #digitalera #chatbot #languagemodel
October 14, 2025 at 6:34 AM
AI Agentはデザインシステムを理解していないという話 "Storybook Design Systems with Agents RFC · storybookjs/ds-mcp-experiment-reshaped · Discussion #1" https://github.com/storybookjs/ds-mcp-experiment-reshaped/discussions/1 #LanguageModel
October 8, 2025 at 1:05 PM
Fuel Your LLM with High-Quality Training Data

Scale smarter. Train faster. Perform better.

Learn more: shorturl.at/BJZIA

#LLM #DataServices #Data #MachineLearning #GenerativeAI #TrainingData #DataAnnotation #LanguageModel #NLP
LLM Training Data Services for Fine-Tuning & RLHF
Boost your AI development with LLM training data services tailored for fine-tuning, RLHF, annotation, and RAG. Get high-quality, domain-specific datasets.
shorturl.at
October 6, 2025 at 6:08 AM
Fuel Your LLM with High-Quality Training Data

Scale smarter. Train faster. Perform better.

Learn more: shorturl.at/BJZIA

#LLM #DataServices #Data #MachineLearning #GenerativeAI #TrainingData #DataAnnotation #LanguageModel #NLP
LLM Training Data Services for Fine-Tuning & RLHF
Boost your AI development with LLM training data services tailored for fine-tuning, RLHF, annotation, and RAG. Get high-quality, domain-specific datasets.
shorturl.at
October 6, 2025 at 6:03 AM
Explore the world of AI and language translation with GPT-4! Discover limitless possibilities and revolutionize communication. #LanguageModel #AI https://fefd.link/uRkqk
October 5, 2025 at 1:52 PM
Towards functional annotation with latent protein languagemodel features [new]
Protein LMs: create functional annotations. Pipeline finds cohesive, reliable features and detects missing annotations.
October 5, 2025 at 5:09 AM
Towards functional annotation with latent protein languagemodel features https://www.biorxiv.org/content/10.1101/2025.10.02.680154v1
October 5, 2025 at 4:47 AM
Towards functional annotation with latent protein languagemodel features https://www.biorxiv.org/content/10.1101/2025.10.02.680154v1
October 5, 2025 at 4:47 AM
LM StudioのAPIを使ったepubの翻訳ツール "sumik5/llm-translate" https://github.com/sumik5/llm-translate/tree/main #translate #LanguageModel
October 4, 2025 at 12:56 AM
October 2, 2025 at 9:19 PM
Explore the world of AI! Discover how OpenAI's GPT-4 revolutionizes language understanding. #GPT4 #LanguageModel #OpenAI https://fefd.link/54ydt
October 2, 2025 at 10:52 AM
#XaniaMonet in Kehlani´s eyes by @billboard.com

The Song-Generator-Company is already in a BIG lawsuit for pirating youtube videos:

www.billboard.com/music/music-...

Just like a "Text" made with a Languagemodel trained on pirated content is NOT yours.

#creative #art @ai.bots.law #aifails #news
Kehlani Slams AI Artist Xania Monet Over $3 Million Record Deal Offer: ‘I Don’t Respect It’
Kehlani slammed AI artist Xania Monet's $3 million record deal, telling followers, "I don't respect it" in a video.
www.billboard.com
October 1, 2025 at 5:56 PM
A 1B-parameter language model boosted data efficiency via latent-thought inference, gaining improvements after three EM cycles without an external teacher model. Read more: https://getnews.me/latent-thought-modeling-improves-data-efficiency-in-lm-pretraining/ #languagemodel #latentthought
October 1, 2025 at 12:56 PM
DiDi‑Instruct speeds language generation up to 64× and reaches a perplexity of 62.2 with just eight NFEs. Training time drops about twenty‑fold versus standard fine‑tuning. https://getnews.me/didi-instruct-boosts-language-generation-speed-by-up-to-64x/ #didiinstruct #languagemodel #ai
October 1, 2025 at 2:19 AM
How “domestic” is a #Victorian novel?
Guhr et al. fine-tune a #LanguageModel to detect implicit domestic spaces – rooms, gardens, even #ships – beyond obvious keywords like 'house' or 'home.' – A new way to read #19th-century #fiction through the lens of #space and study the rise of #domesticity.
September 30, 2025 at 5:39 AM