🎃 Spooky Halloween, folks! Seriously though, you know what haunts me: Climate Change. And today marks the beginning of the 26th UN Climate Change Conference (COP26) — meaning it's been six long years since the Paris one. You know, the famous conference where countries made a covenant to take our impending doom seriously — yet I'm afraid not much has changed.
🌱 So there's a lot that hinges on mitigating the effect of climate change, such as the lives and livelihoods of millions of people that live in low-lying river deltas and island nations or those that live in desertifying or water-stressed regions. As a data scientist in agriculture, I worry how this will lead to crop failures and even more aggressive plant diseases making more people die from hunger and malnutrition. Also, business as usual will cause thousands of ecosystems to collapse from swamps to corals to tundra to rainforests. And along with it, the extinction of up to one-third of the planet's species. The value of this damage is incalculable and, likely, irreversible. To me, the social and environmental dystopia that awaits should be enough to spring world leaders into action.
💸 However, the perhaps more convincing argument to those holding the purse strings is that the global economy will suffer more significant losses if we do nothing. According to Morgan Stanley and International Renewable Energy Agency, the cost of reducing global warming is somewhere between $50 and $131 Trillion, respectively. To put these figures in perspective, the global economy is currently about $85 trillion. By 2050, the global economy could lose anywhere between $ 25 trillion, for a 2°C increase, and $42 trillion, for a 3.2°C increase. These are annual figures, so it would really pay for itself. Not to mention, these are conservative figures from Swiss Re (one of the world's largest reinsurance providers). Researchers from the Imperial College London and London School of Economics have estimates twice as large!
🍭 Anyway, let's hope COP26 leads to meaningful action so that the happy children collecting candies tonight don't have a bleak future contending with the dire consequences of inaction.
Humans have this notion of fairness hardwired from a young age. However, a Machine Learning model's "rules" are not guaranteed to be consistent because we fit models to optimize predictive performance regardless of consistency.
💸 For instance, training data might suggest that banks customers are more likely to default on their loans at a younger age. However, because of less training data and outliers for some ages, the model learns that the probability of defaulting on loans doesn't consistently decrease with age. We call this non-monotonic.
📈 In mathematics, a monotonic function is one that always increases or decreases. And some ML model classes (such as XGBoost) will let you define monotonic constraints. In this #ml example, these constraints would force the model to learn a consistent relationship between loan defaults and age.
⚖️ So why is this fairer? Because similar people should receive equal treatment. The definition of what similar means depends on domain knowledge about a problem. For instance, we might know that higher grades for law students lead to passing the bar exam. So perhaps it's fair to ensure that a law school admissions model consistently favors students with better grades. Another desirable property is that it's easier to explain outcomes for a model with such constraints, and it generalizes better.
🔎 However, if we take a hard look, we'll notice deeper bias. For instance, maybe underprivileged students consistently get lower grades. We can't assume that optimizing predictive performance nor even equalizing outcomes will ensure fairness. Models have the power of changing the future for the better, so why just settle for predicting it based on a lousy past? Constraints serve as guardrails precisely to be leveraged in these cases.
Artificial intelligence and data science must instill trust because good decision-making depends on it, which, in turn, drives better outcomes, reputation, and ultimately adoption.
🤖 So it's a core message of my book that if we are to replace or extend software systems with A.I. systems, we have to guarantee improvements in trustworthiness. And producing trustworthy insights and models is a constant struggle in data science.
⚖️ Interpretable Machine Learning (a.k.a Explainable AI) provides tools to address trust/ethical concerns organized in three levels: 𝗙𝗮𝗶𝗿𝗻𝗲𝘀𝘀, 𝗔𝗰𝗰𝗼𝘂𝗻𝘁𝗮𝗯𝗶𝗹𝗶𝘁𝘆, 𝗮𝗻𝗱 𝗧𝗿𝗮𝗻𝘀𝗽𝗮𝗿𝗲𝗻𝗰𝘆 — collectively known as F.A.T. I like to see these in a pyramid structure because each level depends on the one beneath it. And there are interpretability tools to diagnose problems on each level as well as to fix each problem.
📖 It's an extensive area of active research with hundreds of methods. My book is an introduction with several forays into advanced topics.
It's hard to tell from all the hype, but Artificial Intelligence 𝗶𝘀 𝗯𝗮𝗿𝗲𝗹𝘆 𝗶𝗻 𝗶𝗻𝗳𝗮𝗻𝗰𝘆 👶. But I'm hopeful that we can bring it into maturity.
☔ One of the most significant issues Machine Learning projects face is that models are ill-equipped to weather changing, adversarial, and 𝘂𝗻𝗲𝘅𝗽𝗲𝗰𝘁𝗲𝗱 𝗱𝗮𝘁𝗮 𝗰𝗼𝗻𝗱𝗶𝘁𝗶𝗼𝗻𝘀, much like planes facing storms and turbulence. But aircraft are robustly built and can overcome severe conditions both automatically and guided by experienced pilots. On the other hand, we know models must generalize well, but this proves to be an elusive property.
🎛️ Ever since I wrote my book, I've been asked many times why I'm passionate about 𝗜𝗻𝘁𝗲𝗿𝗽𝗿𝗲𝘁𝗮𝗯𝗹𝗲 𝗠𝗮𝗰𝗵𝗶𝗻𝗲 𝗟𝗲𝗮𝗿𝗻𝗶𝗻𝗴. I've responded that it's the instrument panel to pilot machine learning even in the worst conditions, from unfair to uncertain outcomes. So why wouldn't I prefer to have a complete instrument panel available? But, on the other hand, using predictive performance alone is like piloting with a single instrument!
✈️ Currently, flying is the safest mode of transportation. But for A.I., there is still a long way to go. For starters, we will need better no-code AutoML with human-in-the-loop and Interpretable M.L. built-in — like cockpits for Machine Learning engineers. And methods that automatically audit and test models, much like commercial planes, undergo strict maintenance regimens. And given what I've seen currently being built by Auto ML, MLOps, and XAI startups and researchers, it seems like it's heading in this direction, so I have reasons to be hopeful that for most commercial use cases, A.I. someday will be the 𝘀𝗮𝗳𝗲𝘀𝘁 𝗺𝗼𝗱𝗲 𝗼𝗳 𝗱𝗲𝗰𝗶𝘀𝗶𝗼𝗻-𝗺𝗮𝗸𝗶𝗻𝗴!
🥏 Recently, I found this box of frisbees in my parent's basement, and it's what's left of my biggest failure — a search engine #startup. 𝗙𝗮𝗶𝗹𝘂𝗿𝗲 𝘀𝗼𝘂𝗻𝗱𝘀 𝗵𝗮𝗿𝘀𝗵, but we learn by trial and error, so every mistake is an opportunity for growth.
📊 One of the biggest lessons I learned was technical, and it had to do with the importance of #analytics and 𝗱𝗲𝗯𝘂𝗴 𝗮𝗹𝗴𝗼𝗿𝗶𝘁𝗵𝗺𝘀 for points of failure. It was then I realized that Machine Learning had a problem. After all, how do you debug ML models? This is how in 2017, I first stumbled upon Interpretable ML / Explainable AI research. Fast forward to 2020, and I was writing a book about it! And I spoke about this journey to San-Francisco-based A.I. startup entrepreneurs and workers.
💪 𝐼𝑛 𝐶𝑜𝑛𝑐𝑙𝑢𝑠𝑖𝑜𝑛: the frisbees may have been the only tangible items, but my failure left behind stories, ideas, lessons, and a brand new perspective — that has only made me 𝘀𝘁𝗿𝗼𝗻𝗴𝗲𝗿! As for the frisbees, they will find a new home with goodwill.
I learned to program on this computer — I was a child during the '80s 🤓. It had a 4.77 MHz CPU, 256 KB RAM, monochrome display, and no hard drive, so you had to be creative to overcome 𝗿𝗲𝘀𝗼𝘂𝗿𝗰𝗲 𝗰𝗼𝗻𝘀𝘁𝗿𝗮𝗶𝗻𝘁𝘀 — not to mention exercise patience!
We are 𝘀𝗼 𝘀𝗽𝗼𝗶𝗹𝗲𝗱 these days! To put it in context, most smartphones 📱 have over 16 thousand times the RAM and more storage than would have fit in a room in the 80s. Add that to cheap, limitless cloud storage. I am not complaining.. That is great! However, I wonder how much does resource constraints foster software innovation — and optimal code.
Today, trillion-parameters deep learning 🤖 models are pushing the envelope. Still, at the same time, it seems illogical that they represent the most 𝗲𝗳𝗳𝗶𝗰𝗶𝗲𝗻𝘁 𝘀𝗼𝗹𝘂𝘁𝗶𝗼𝗻 grounded in, for instance, biology, causal understanding of the world, or statistics. So before ushering in the age of quantum computing, I'm hoping we hit some resource limitations to focus more energy on more creative and intuitive solutions — not to mention cost-effective.
What do you think? How much does an abundance of resources hinder or enable creative solutions?
🇨🇷 7 years ago, I had a fantastic 4-day journey trekking through the 𝗖𝗼𝘀𝘁𝗮 𝗥𝗶𝗰𝗮𝗻 𝗿𝗮𝗶𝗻𝗳𝗼𝗿𝗲𝘀𝘁. On the 1st day, we had to cross a wild river with a metal basket hanging on a rusty rope. And I thought to myself, "what the hell have I gotten into?!".
🐒 On that journey, I saw 𝗺𝗮𝗻𝘆 𝘀𝗽𝗲𝗰𝗶𝗲𝘀 of wildlife. I slept smelling the moss on the bark and wet ferns. And I woke up every morning to a majestic orchestra of birds, insects, monkeys, and frogs. It's also hard to realize the sheer scale of a rainforest when you are in it. On peaks, we could see the many green valleys we had crossed with Ceiba trees towering 17 stories high over the canopy!
🌎 We only have 36% of rainforests left. When I was born it was well over 50%. Today is #WorldRainforestDay and I thought I’d share a story of why I care. In #DataScience, we think 𝘧𝘢𝘤𝘵𝘴 & 𝘧𝘪𝘨𝘶𝘳𝘦𝘴 alone are convincing. But often it's the 𝘭𝘪𝘷𝘦𝘥 𝘦𝘹𝘱𝘦𝘳𝘪𝘦𝘯𝘤𝘦 & 𝘦𝘮𝘰𝘵𝘪𝘰𝘯𝘴 that come with them that make things matter to us. I don't regret crossing the river on the basket because the journey the followed was life-changing. If I was an environmentalist before because of the facts I knew, now I had more conviction than ever that #nature had to be preserved for future generations!
Today is 𝐖𝐨𝐫𝐥𝐝 𝐅𝐨𝐨𝐝 𝐒𝐚𝐟𝐞𝐭𝐲 𝐃𝐚𝐲. For me, it's a day of reflection.
🦠 After all, 𝗖𝗢𝗩𝗜𝗗𝟭𝟵 had a food safety-related genesis. Natural disasters and clearing land for urbanization + agriculture pushes wildlife closer to human settlements, which fuel pandemic risk.
🌽 Food safety is essential, no doubt, but it's intrinsically related to 𝗳𝗼𝗼𝗱 𝘀𝗲𝗰𝘂𝗿𝗶𝘁𝘆 — and this is what worries me the most. There's a need to feed another 2 billion mouths by 2050 and double food production to that end. So, as a data scientist in agriculture, I'm inspired to make my tiny contribution to improving food security.
🌎 However, 𝗰𝗹𝗶𝗺𝗮𝘁𝗲 𝗰𝗵𝗮𝗻𝗴𝗲 can make our food production goals nearly impossible. Under a high-emission scenario, by 2050, a huge swath of the United States will suffer from a sizable decline in crop yields. However, this is offset by the fact that other areas of the country will experience an increase. Other countries won't be that lucky since they are entirely vulnerable given so many land challenges: desertification, land degradation, climate change adaptation, undernourishment, biodiversity, groundwater stress, and water quality (see IPCC for details). It's an existential threat to humanity, and we have only a few years to reverse this trajectory.
When discussing human judgment and, by extension, algorithmic decisions, we are used to talking about 𝐛𝐢𝐚𝐬, but what about 𝐧𝐨𝐢𝐬𝐞?
🎯 Nobel Laureate ᴅᴀɴɪᴇʟ ᴋᴀʜɴᴇᴍᴀɴ and co-authors make a case for why we should pay close attention to it in their new book 𝑁𝑜𝑖𝑠𝑒: 𝐴 𝐹𝑙𝑎𝑤 𝑖𝑛 𝐻𝑢𝑚𝑎𝑛 𝐽𝑢𝑑𝑔𝑒𝑚𝑒𝑛𝑡. It has some compelling stories to underpin how widespread the problem is in business and government with succinct illustrations. For instance, I love the target illustration and the error decompositions.
📢 The book covers group dynamics such as information cascades, social pressure, group polarization as amplifiers of noise, and some cognitive #biases to boot. Lastly, it outlines noise mitigation strategies with decision hygiene, decision observers, and noise audits, which were BY FAR the biggest takeaways for me.
😒 However, if you are already familiar with the topic, the book will likely disappoint (at least a little). It can feel very repetitive and not getting into enough depth, and its entanglement with bias means it keeps referring to concepts covered in 𝑇ℎ𝑖𝑛𝑘𝑖𝑛𝑔 𝐹𝑎𝑠𝑡 𝑎𝑛𝑑 𝑆𝑙𝑜𝑤, as it was some long-lost final chapter. I still enjoyed it, regardless.
Have you read it? Do you want to?
🎲 𝐂𝐡𝐚𝐧𝐜𝐞 & 📈 𝐃𝐞𝐜𝐢𝐬𝐢𝐨𝐧-𝐌𝐚𝐤𝐢𝐧𝐠 — Long before I called myself a data scientist, I helped build a backend website for sports betting on my very first job. After that, for about a decade, I improved the user experience for gambling sites of all kinds. As it turns out, the first data I engaged with professionally taught me a lot about human nature.
Throughout human history, we have been fascinated with chance. The first known tools used to this end were knucklebones in ancient Sumer, either for fortune-telling or games of chance. Better tools have been invented since, like dice, playing cards, and more recently, random number generators (RNGs). However, now we wield randomness for business/scientific purposes and not just mysticism/entertainment. In fact, the most powerful #MachineLearning methods depend on RNGs.
I recently read the book 𝘛𝘩𝘦 𝘋𝘳𝘶𝘯𝘬𝘢𝘳𝘥𝘴 𝘞𝘢𝘭𝘬, which made me reflect on what drew me to the discipline. We are surrounded by randomness, but humans want to be in control, often attributing skill to successful random events (ɢᴀᴍʙʟᴇʀ'ꜱ ꜰᴀʟʟᴀᴄʏ), and lack thereof otherwise. #Data can improve decisions by separating the signal from the noise and tracing outcomes to plausible causes. This possibility is what inspires my journey! What's yours?