<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Calibrated Intuition]]></title><description><![CDATA[A space to develop frameworks of how to think about innovation and how technology will shape development in Africa. I also write about deep learning and AI research.]]></description><link>https://www.victorakinwande.com</link><image><url>https://substackcdn.com/image/fetch/$s_!vy5x!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa841dcb9-f42d-4381-9238-0f309b5040c5_544x544.png</url><title>Calibrated Intuition</title><link>https://www.victorakinwande.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 05 Apr 2026 00:35:28 GMT</lastBuildDate><atom:link href="https://www.victorakinwande.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Victor Akinwande]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[victorakinwande@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[victorakinwande@substack.com]]></itunes:email><itunes:name><![CDATA[Victor Akinwande]]></itunes:name></itunes:owner><itunes:author><![CDATA[Victor Akinwande]]></itunes:author><googleplay:owner><![CDATA[victorakinwande@substack.com]]></googleplay:owner><googleplay:email><![CDATA[victorakinwande@substack.com]]></googleplay:email><googleplay:author><![CDATA[Victor Akinwande]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Understanding prompt engineering may not require rethinking generalisation]]></title><description><![CDATA[Tldr: When learning over prompts, classic PAC-Bayes bounds are extremely tight; more so than any previous PAC-Bayes analysis of deep networks.]]></description><link>https://www.victorakinwande.com/p/understanding-prompt-engineering</link><guid isPermaLink="false">https://www.victorakinwande.com/p/understanding-prompt-engineering</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Thu, 23 Nov 2023 02:22:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!U863!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Tldr: </strong>When learning over prompts, <em>classic</em> PAC-Bayes bounds are extremely tight; more so than any previous PAC-Bayes analysis of deep networks.</p><p><strong>Zero-shot performance of visual-language models</strong></p><p>An emerging paradigm in deep learning is to use large pre-trained models such as <a href="https://arxiv.org/abs/2103.00020">CLIP</a> or <a href="https://arxiv.org/abs/2102.05918">ALIGN</a> as feature extractors and provide weak supervision for a downstream target task via prompts, which are text descriptions of the desired tasks or concept to classify. This performs remarkably well &#8211; up to 75% accuracy on ImageNet. This has led to handcrafting of prompts or prompt engineering. When practitioners attempt to engineer prompts for a dataset - they often are not concerned with overfitting to the test set. In recent work: <a href="https://arxiv.org/abs/2310.03957">https://arxiv.org/abs/2310.03957</a> we illustrate that in-fact you can&#8217;t overfit.</p><p><strong>Observation:</strong> Given how small the vocabulary sizes and context lengths of such visual-language models are, the hypothesis class is relatively small compared to what is often studied in generalisation of neural networks. For example, the VC dimension of neural networks is typically lower bounded by the number of parameters - in direct conflict with the paradigm of large models.</p><p><strong>Numerical non-vacuous bounds in deep learning</strong></p><p>Whereas obtaining non-vacuous bounds for deep neural networks still remains a technical challenge when done on the weight space. Given a fixed image encoder, the text encoder of CLIP, for instance, is an injective map of the prompts to the weights of a linear classifier. One could evaluate classic PAC-Bayes bounds on the prompts. I show here the classic PAC-Bayes bound [<a href="https://www.semanticscholar.org/paper/PAC-Bayesian-model-averaging-McAllester/609be06b5a91ffb8c35abd1636cd2431ff32435d">McAllester 1999</a>].</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JU-Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JU-Q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 424w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 848w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 1272w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JU-Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png" width="464" height="76.27397260273973" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/df85b458-42ca-4344-9623-c252b0148be0_1022x168.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:168,&quot;width&quot;:1022,&quot;resizeWidth&quot;:464,&quot;bytes&quot;:27387,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JU-Q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 424w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 848w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 1272w, https://substackcdn.com/image/fetch/$s_!JU-Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf85b458-42ca-4344-9623-c252b0148be0_1022x168.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>The dominant term is the KL divergence of the posterior from the prior. CLIP for instance has a vocabulary of about 50,000 tokens and a context length of 77. A uniform prior means that KL term = 77 &#8727; log&#8289;(49408). Note we're now operating in the discrete space of language tokens. This implies on a dataset like CIFAR-10 with 50000 training examples, 10 classes - we have a bound of ~30% on the risk any prompt-based classifier with 99% confidence. This is already much tighter than bounds we typically get for neural networks.</p><p>Clearly, we can do better with a more informative prior - e.g. a language model. Here&#8217;s a plot of the PAC-Bayes bound using the LLamA-7B as the prior. The way to interpret the plot is if the true risk according to the bound, is equal to the empirical risk - then all the points will be on the dotted line.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U863!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U863!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 424w, https://substackcdn.com/image/fetch/$s_!U863!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 848w, https://substackcdn.com/image/fetch/$s_!U863!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 1272w, https://substackcdn.com/image/fetch/$s_!U863!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U863!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png" width="282" height="243.7037037037037" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fc08317a-ad35-4941-9692-7e35a14f4289_486x420.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:420,&quot;width&quot;:486,&quot;resizeWidth&quot;:282,&quot;bytes&quot;:57331,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!U863!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 424w, https://substackcdn.com/image/fetch/$s_!U863!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 848w, https://substackcdn.com/image/fetch/$s_!U863!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 1272w, https://substackcdn.com/image/fetch/$s_!U863!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffc08317a-ad35-4941-9692-7e35a14f4289_486x420.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This tells us if we simply use the prompt - a photo of a {class_name} for all the 1000 classes in ImageNet - our true risk will be no more than ~30% with high probability. Cool!</p><p>What&#8217;s the catch? Well these bounds rely on the fact that pre-trained visual-language models already contain some hypothesis class that will perform well on the training set - and the generalisation of the underlying models themselves is not being analysed but what this does provide is solid evidence for why the wide-spread practice of prompt-engineering is as principled as you can get from a statistical learning theory perspective.</p>]]></content:encoded></item><item><title><![CDATA[That Neem Tree]]></title><description><![CDATA[There, where paths crossed and stories unfolded, Stood the ancient neem, witness to the ages.]]></description><link>https://www.victorakinwande.com/p/that-neem-tree</link><guid isPermaLink="false">https://www.victorakinwande.com/p/that-neem-tree</guid><dc:creator><![CDATA[Victor Akinwande]]></dc:creator><pubDate>Wed, 22 Nov 2023 22:48:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vy5x!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa841dcb9-f42d-4381-9238-0f309b5040c5_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><code>There, where paths crossed and stories unfolded, Stood the ancient neem, witness to the ages. Its branches, like weathered arms, reached skyward, While its roots, hidden, grasped the earth&#8217;s secrets.</code></p><p><code>There, when the first of them came along, Curiously watching, they measured, planned and built. Taking no notice of your gaze, they occupied, Steel and concrete sprouting beneath your boughs.</code></p><p><code>You, silent sentinel, watched as time danced, All was prosperous and flourished, under your watchful eye. Leaves rustled with whispers of days long gone, And bark etched with memories, bore silent testimony.</code></p><p><code>In the cool shade of your wisdom, they laughed and cried, Unaware of the eyes that have seen empires rise and fall. Should you have acknowledged their fleeting presence? Their good fortune, their sorrows, their mortal worries?</code></p><p><code>Or should you stand, stoic, as you always have, A symbol of resilience in the face of fleeting endeavors, A reminder of the enduring spirit of nature, Against the ever-changing tapestry of human toil?</code></p><p><code>In your silence, there lies a profound understanding, Of cycles of life, of the transient and the eternal. That neem tree, under your boughs, lies the truth - Of nature&#8217;s quiet endurance amidst human ambition.</code></p><p></p><p>- Co-authored with GPT-4</p>]]></content:encoded></item><item><title><![CDATA[Introducing Calibrated Intuition]]></title><description><![CDATA[.]]></description><link>https://www.victorakinwande.com/p/coming-soon</link><guid isPermaLink="false">https://www.victorakinwande.com/p/coming-soon</guid><pubDate>Tue, 14 Jul 2020 18:33:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vy5x!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa841dcb9-f42d-4381-9238-0f309b5040c5_544x544.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to Calibrated Intuition by me, Victor Akinwande. I'm interested in Technology, AI, Cloud, &amp; Energy Policy. </p><p>Calibrated intuition is a space to develop frameworks of how to think about innovation and how technology will shape development in Africa. I will also write technical articles related to my research in AI and deep learning.</p><p>Sign up now and <a href="https://www.victorakinwande.com/p/coming-soon?utm_source=substack&utm_medium=email&utm_content=share&action=share">tell your friends</a>!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.victorakinwande.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.victorakinwande.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item></channel></rss>