βοΈ The PubMed Open-Access (OA) subset shares a metadata for 35 Million articles. Suddenly, the existing article parser represents a Hugging Face dataset that was supported up until 2024. ncbi/pubmed Moreover, the pubmed data represent a compressed XLM which is beneficial for efficiency but limits processing technique application.
π’ To bridge this gap, excited to share pubmed_articles_iter project, which bridges this gap by providing: βοΈ 1. Downloader for the raw files βοΈ 2. No-string iterator over pubmed articles, utilized for converting them into JSON.
So far, BioASQ organizers as CLEF-2025 reveal the complete leaderboard of other submissions (see image below).
Our distil-tuned Qwen2.5-0.5B (BU-Team) has been ofically ranked as the second-best performing system in French! π«π· We also investigate the strongest recall of key aspects among all participants β demonstrating the value of adopted fine-tuning strategy.
π’ For those who interested in adopting streaming with bare minimum dependencies and setting up a GenAI powered demo in Web, this post might be relevant. Streaming support is inevitable for running local or remote models. Delighted to share the first part of the tutorial.
From which you will learn how to: βοΈ Use pure JS for fetching streaming from the specific provider (Replicate) βοΈ Use pure JS with custom proxy streaming provider (FastAPI)
β¨ TLDR: We review POST-based approaches for fetching data readers and adopting data parsers. Using FastAPI as a proxy, we explain how to take control over transferred data.
π’ For those who planning to start a PhD or research in the UK π¬π§ (including AI field in particular) but facing ATAS (Academic Technology Approval Scheme) issues. Excited to share the ultimate guide for dealing with ATAS refusals and how to write effective rebuttal letters.
π From the video you will find: 1. Why appealing an ATAS decision matters even if your visa is approved 2. Which docments to use in understanding the principles behind sponsorship decisions 3. Key tips for proper rebuttal letter structuring
longer context doesn't generate better responses. it can even hurt your llm/agent. 1M context window doesn't automatically make models smarter as it's not about the size; it's how you use it.
here are 4 types of context failure and why each one happens:
1. context poisoning: if hallucination finds its way into your context, the agent will rely on that false information to make its future moves. for example if the agent hallucinates about the "task description", all of its planning to solve the task would also be corrupt.
2. context distraction: when the context becomes too bloated, the model focuses too much on it rather than come up with novel ideas or to follow what it has learned during training. as Gemini 2.5 Pro technical report points out, as context grows significantly from 100K tokens, "the agent showed a tendency toward favoring repeating actions from its vast history rather than synthesizing novel plans".
3. context confusion: everyone lost it when MCPs became popular, it seemed like AGI was achieved. I suspected there is something wrong and there was: it's not just about providing tools, bloating the context with tool use derails the model from selecting the right one! even if you can fit all your tool metadata in the context, as their number grows, the model gets confused over which one to pick.
4. Context Clash: if you exchange conversation with a model step by step and provide information as you go along, chances are you get worse performance rather than providing all the useful information at once. one the model's context fills with wrong information, it's more difficult to guide it to embrace the right info. agents pull information from tools, documents, user queries, etc. and there is a chance that some of these information contradict each other, and it's not good new for agentic applications.