Posts

Competing to Collaborate: How Federated Learning Is Quietly Reshaping Pharma's AI Race

Image
There's an old pharmaceutical paradox: the data that would most accelerate drug discovery is precisely the data companies will never share.  Every major player in the industry has a mountain of data on compounds tested, assay results, and clinical observations – all of which could be a treasure trove for machine learning models. However, sharing this data with a competitor? Out of the question. Federated Learning (FL) is a technology that's silently eliminating this paradox. It allows drug discovery leaders to collaborate on machine learning model development without sharing data with their competitors – not even a peek. And 2024 and 2025 saw this technology go from proof-of-concept to production. Federated Learning (FL) is the technology quietly dismantling that paradox. It lets companies collaborate on AI model training without ever moving, exposing, or even glimpsing each other's underlying data. And in 2024–2025, the industry moved from proof-of-concept to real-world...

AI Basics series : All LLMs are FMs, but not all FMs are LLMs

Image
All LLMs are foundation models, but not all foundation models are LLMs.  Let's break it down to clarify: 1. All LLMs are foundation models: Foundation Models (FMs): These are large, pre-trained models trained on vast amounts of unlabelled data at scale. They are designed to be "foundational" because they can be adapted (fine-tuned) for a wide range of downstream tasks without needing to be trained from scratch for each specific application. Their pre-training allows them to learn general representations and patterns across the data. Large Language Models (LLMs): These are a specific type of foundation model that is specialized in processing and generating human language. They are trained on massive text datasets and excel at tasks like text generation, translation, summarization, and question answering. Since LLMs fit the definition of a foundation model (large, pre-trained on vast data, adaptable), every LLM is inherently a foundation model. 2. Not all foundation mode...

Prompt Engineering: Frequently Asked Questions

Image
What is prompt engineering? Prompt engineering is the art and science of crafting effective inputs (prompts) for large language models (LLMs) like ChatGPT and Claude to elicit desired outputs. It involves understanding the model's capabilities, limitations, and how it interprets language to guide it towards generating accurate, relevant, and creative responses. Why is it called "engineering"? The term "engineering" emphasizes the iterative and systematic approach involved in prompt crafting. It's not simply writing a single sentence; it requires experimentation, refinement, and a deep understanding of how to interact with the model to achieve specific goals. This often involves testing multiple prompts, analyzing outputs, and adjusting the language, format, and structure until the desired results are consistently achieved. How do I write a good prompt? A good prompt is clear, concise, and specific, providing enough context for the model to understand the tas...

FAQ's on Optical Networking Technologies

Image
1. What is DWDM and how does it increase network capacity? DWDM stands for Dense Wavelength-Division Multiplexing. It's a technology that increases network capacity by transmitting multiple data streams on different wavelengths (colors) of light through a single optical fiber. This allows for a significant increase in bandwidth without laying new fiber, making it a cost-effective way to enhance network capacity. 2. What are the key differences between PON and AON networks? PON (Passive Optical Network) and AON (Active Optical Network) are both fiber optic network architectures, but they differ in how they transmit data: PON uses passive optical splitters to divide an incoming optical signal from the OLT (Optical Line Terminal) to multiple ONUs (Optical Network Units). This makes PON more cost-effective but limits bandwidth per user as it's shared. AON uses active components like repeaters to amplify and distribute the signal, providing dedicated bandwidth to each user. AON of...

Tech Podcast Brief - How HPC & AI are Changing DC Networks

Image
Source: Heavy Networking podcast episode with Rob Sherwood, discussing the impact of High-Performance Computing (HPC) and Artificial Intelligence (AI) on data center network design. Main Themes: The unique demands of AI training workloads necessitate dedicated network infrastructure. Traditional networking assumptions like oversubscription and best-effort delivery do not apply to HPC and AI. Bandwidth, power, and cooling are major challenges that require innovative solutions. The network interface card (NIC) architecture is evolving to address these challenges, with a shift towards smarter NICs, RDMA, and even optical interconnects. Most Important Ideas/Facts: 1. Collective Communication: AI training, especially building large language models (LLMs), relies on collective communication operations like "all-reduce", where data is exchanged and processed simultaneously across all nodes. Traditional unicast-based networks are ill-suited for this, as they lead to congestion, pac...

The Quantum Computing Revolution

Image
The emergence of quantum computers signifies a transformative leap in computational technology, promising advancements that could surpass the profound impact of traditional digital computers. Quantum computing operates on the principles of quantum mechanics, utilizing the behavior of atoms and particles to perform calculations far beyond the capabilities of conventional systems. This post is the excerpts from the video talk by Dr. Michio Kaku available on BIG THINK youtube channel. The Race for Quantum Computing The competition to perfect quantum computers involves all major technology companies and national governments. Non-participation in this race could lead to Silicon Valley becoming analogous to the Rust Belt. Security agencies, including the FBI and CIA, are closely monitoring developments in quantum computing due to its potential to crack codes based on digital technology. Historical Context of Computing Three stages of computing: Analog Computers: Originated over 2,000 yea...

InfiniBand vs. Ethernet vs. Ultra Ethernet

The high-performance computing (HPC) and data center realms are witnessing a dynamic interplay between InfiniBand, Ethernet, and the emerging Ultra Ethernet. Each technology offers distinct strengths and caters to specific application demands. InfiniBand Historically, InfiniBand has been the preferred choice for HPC environments due to its exceptionally low latency and high bandwidth. It was architected for data-centric workloads, providing robust error correction, congestion control, and efficient data transfers. While offering unparalleled performance, InfiniBand has traditionally been associated with higher costs and a less extensive ecosystem compared to Ethernet. Ethernet Ethernet, once primarily a technology for local area networks (LANs), has undergone a remarkable evolution. Advancements in Ethernet technology, such as RDMA (Remote Direct Memory Access) and low-latency switches, have significantly bridged the performance gap with InfiniBand. Its ubiquity, cost-effectiveness, an...