#probabilistic programming
Explore tagged Tumblr posts
Note
Yeah come play our randomizer!
(Though these days I mostly work on the Ori 1 Randomizer, which is also an incredible software abomination even if the seed specification format isn't a fully fledged programming language. It has great new player tools! Come try it! If you played Ori once you can play our randomizer, I promise!!)
How do you *accidentally* make a programming language?
Oh, it's easy! You make a randomizer for a game, because you're doing any% development, you set up the seed file format such that each line of the file defines an event listener for a value change of an uberstate (which is an entry of the game's built-in serialization system for arbitrary data that should persiste when saved).
You do this because it's a fast hack that lets you trigger pickup grants on item finds, since each item find always will correspond with an uberstate change. This works great! You smile happily and move on.
There's a small but dedicated subgroup of users who like using your randomizer as a canvas! They make what are called "plandomizer seeds" ("plandos" for short), which are seed files that have been hand-written specifically to give anyone playing them a specific curated set of experiences, instead of something random. These have a long history in your community, in part because you threw them a few bones when developing your last randomizer, and they are eager to see what they can do in this brave new world.
A thing they pick up on quickly is that there are uberstates for lots more things than just item finds! They can make it so that you find double jump when you break a specific wall, or even when you go into an area for the first time and the big splash text plays. Everyone agrees that this is neat.
It is in large part for the plando authors' sake that you allow multiple line entries for the same uberstate that specify different actions - you have the actions run in order. This was a feature that was hacked into the last randomizer you built later, so you're glad to be supporting it at a lower level. They love it! It lets them put multiple items at individual locations. You smile and move on.
Over time, you add more action types besides just item grants! Printing out messages to your players is a great one for plando authors, and is again a feature you had last time. At some point you add a bunch for interacting with player health and energy, because it'd be easy. An action that teleports the player to a specific place. An action that equips a skill to the player's active skill bar. An action that removes a skill or ability.
Then, you get the brilliant idea that it'd be great if actions could modify uberstates directly. Uberstates control lots of things! What if breaking door 1 caused door 2 to break, so you didn't have to open both up at once? What if breaking door 2 caused door 1 to respawn, and vice versa, so you could only go through 1 at a time? Wouldn't that be wonderful? You test this change in some simple cases, and deploy it without expecting people to do too much with it.
Your plando authors quickly realize that when actions modify uberstates, the changes they make can trigger other actions, as long as there are lines in their files that listen for those. This excites them, and seems basically fine to you, though you do as an afterthought add an optional parameter to your uberstate modification action that can be used to suppress the uberstate change detector, since some cases don't actually want that behavior.
(At some point during all of this, the plando authors start hunting through the base game and cataloging unused uberstates, to be used as arbitrary variables for their nefarious purposes. You weren't expecting that! Rather than making them hunt down and use a bunch of random uberstates for data storage, you sigh and add a bunch of explicitly-unused ones for them to play with instead.)
Then, your most arcane plando magician posts a guide on how to use the existing systems to set up control flow. It leverages the fact that setting an uberstate to a value it already has does not trigger the event listener for that uberstate, so execution can branch based on whether or not a state has been set to a specific value or not!
Filled with a confused mixture of pride and fear, you decide that maybe you should provide some kind of native control flow structure that isn't that? And because you're doing a lot of this development underslept and a bit past your personal Balmer peak, the first idea that you have and implement is conditional stops, which are actions that halt processing of a multiple-action-chain if an uberstate is [less than, equal to, greater than] a given value.
The next day, you realize that your seed specification format now can, while executing an action chain, read from memory, write to memory, branch based on what it finds in memory, and loop. It can simulate a turing machine, using the uberstates as tape. You set out to create a format by which your seed generator could talk to your client mod, and have ended up with a turing complete programming language. You laugh, and laugh, and laugh.
#programming#my projects#randomizers#ori and the blind forest#ori and the will of the wisps#ori rando#we just put out a huge patch for april fools and it's been amazing#4 entire new features dropped by surprise#and they're all dumb!#come try our randomizer randomizer#you will certainly not regret trying the randomizer randomizer#(...you will probabilistically regret trying the randomizer randomizer)
2K notes
·
View notes
Text
Solar is a market for (financial) lemons

There are only four more days left in my Kickstarter for the audiobook of The Bezzle, the sequel to Red Team Blues, narrated by @wilwheaton! You can pre-order the audiobook and ebook, DRM free, as well as the hardcover, signed or unsigned. There's also bundles with Red Team Blues in ebook, audio or paperback.
Rooftop solar is the future, but it's also a scam. It didn't have to be, but America decided that the best way to roll out distributed, resilient, clean and renewable energy was to let Wall Street run the show. They turned it into a scam, and now it's in terrible trouble. which means we are in terrible trouble.
There's a (superficial) good case for turning markets loose on the problem of financing the rollout of an entirely new kind of energy provision across a large and heterogeneous nation. As capitalism's champions (and apologists) have observed since the days of Adam Smith and David Ricardo, markets harness together the work of thousands or even millions of strangers in pursuit of a common goal, without all those people having to agree on a single approach or plan of action. Merely dangle the incentive of profit before the market's teeming participants and they will align themselves towards it, like iron filings all snapping into formation towards a magnet.
But markets have a problem: they are prone to "reward hacking." This is a term from AI research: tell your AI that you want it to do something, and it will find the fastest and most efficient way of doing it, even if that method is one that actually destroys the reason you were pursuing the goal in the first place.
https://learn.microsoft.com/en-us/security/engineering/failure-modes-in-machine-learning
For example: if you use an AI to come up with a Roomba that doesn't bang into furniture, you might tell that Roomba to avoid collisions. However, the Roomba is only designed to register collisions with its front-facing sensor. Turn the Roomba loose and it will quickly hit on the tactic of racing around the room in reverse, banging into all your furniture repeatedly, while never registering a single collision:
https://www.schneier.com/blog/archives/2021/04/when-ais-start-hacking.html
This is sometimes called the "alignment problem." High-speed, probabilistic systems that can't be fully predicted in advance can very quickly run off the rails. It's an idea that pre-dates AI, of course – think of the Sorcerer's Apprentice. But AI produces these perverse outcomes at scale…and so does capitalism.
Many sf writers have observed the odd phenomenon of corporate AI executives spinning bad sci-fi scenarios about their AIs inadvertently destroying the human race by spinning off in some kind of paperclip-maximizing reward-hack that reduces the whole planet to grey goo in order to make more paperclips. This idea is very implausible (to say the least), but the fact that so many corporate leaders are obsessed with autonomous systems reward-hacking their way into catastrophe tells us something about corporate executives, even if it has no predictive value for understanding the future of technology.
Both Ted Chiang and Charlie Stross have theorized that the source of these anxieties isn't AI – it's corporations. Corporations are these equilibrium-seeking complex machines that can't be programmed, only prompted. CEOs know that they don't actually run their companies, and it haunts them, because while they can decompose a company into all its constituent elements – capital, labor, procedures – they can't get this model-train set to go around the loop:
https://pluralistic.net/2023/03/09/autocomplete-worshippers/#the-real-ai-was-the-corporations-that-we-fought-along-the-way
Stross calls corporations "Slow AI," a pernicious artificial life-form that acts like a pedantic genie, always on the hunt for ways to destroy you while still strictly following your directions. Markets are an extremely reliable way to find the most awful alignment problems – but by the time they've surfaced them, they've also destroyed the thing you were hoping to improve with your market mechanism.
Which brings me back to solar, as practiced in America. In a long Time feature, Alana Semuels describes the waves of bankruptcies, revealed frauds, and even confiscation of homeowners' houses arising from a decade of financialized solar:
https://time.com/6565415/rooftop-solar-industry-collapse/
The problem starts with a pretty common finance puzzle: solar pays off big over its lifespan, saving the homeowner money and insulating them from price-shocks, emergency power outages, and other horrors. But solar requires a large upfront investment, which many homeowners can't afford to make. To resolve this, the finance industry extends credit to homeowners (lets them borrow money) and gets paid back out of the savings the homeowner realizes over the years to come.
But of course, this requires a lot of capital, and homeowners still might not see the wisdom of paying even some of the price of solar and taking on debt for a benefit they won't even realize until the whole debt is paid off. So the government moved in to tinker with the markets, injecting prompts into the slow AIs to see if it could coax the system into producing a faster solar rollout – say, one that didn't have to rely on waves of deadly power-outages during storms, heatwaves, fires, etc, to convince homeowners to get on board because they'd have experienced the pain of sitting through those disasters in the dark.
The government created subsidies – tax credits, direct cash, and mixes thereof – in the expectation that Wall Street would see all these credits and subsidies that everyday people were entitled to and go on the hunt for them. And they did! Armies of fast-talking sales-reps fanned out across America, ringing dooorbells and sticking fliers in mailboxes, and lying like hell about how your new solar roof was gonna work out for you.
These hustlers tricked old and vulnerable people into signing up for arrangements that saw them saddled with ballooning debt payments (after a honeymoon period at a super-low teaser rate), backstopped by liens on their houses, which meant that missing a payment could mean losing your home. They underprovisioned the solar that they installed, leaving homeowners with sky-high electrical bills on top of those debt payments.
If this sounds familiar, it's because it shares a lot of DNA with the subprime housing bubble, where fast-talking salesmen conned vulnerable people into taking out predatory mortgages with sky-high rates that kicked in after a honeymoon period, promising buyers that the rising value of housing would offset any losses from that high rate.
These fraudsters knew they were acquiring toxic assets, but it didn't matter, because they were bundling up those assets into "collateralized debt obligations" – exotic black-box "derivatives" that could be sold onto pension funds, retail investors, and other suckers.
This is likewise true of solar, where the tax-credits, subsidies and other income streams that these new solar installations offgassed were captured and turned into bonds that were sold into the financial markets, producing an insatiable demand for more rooftop solar installations, and that meant lots more fraud.
Which brings us to today, where homeowners across America are waking up to discover that their power bills have gone up thanks to their solar arrays, even as the giant, financialized solar firms that supplied them are teetering on the edge of bankruptcy, thanks to waves of defaults. Meanwhile, all those bonds that were created from solar installations are ticking timebombs, sitting on institutions' balance-sheets, waiting to go blooie once the defaults cross some unpredictable threshold.
Markets are very efficient at mobilizing capital for growth opportunities. America has a lot of rooftop solar. But 70% of that solar isn't owned by the homeowner – it's owned by a solar company, which is to say, "a finance company that happens to sell solar":
https://www.utilitydive.com/news/solarcity-maintains-34-residential-solar-market-share-in-1h-2015/406552/
And markets are very efficient at reward hacking. The point of any market is to multiply capital. If the only way to multiply the capital is through building solar, then you get solar. But the finance sector specializes in making the capital multiply as much as possible while doing as little as possible on the solar front. Huge chunks of those federal subsidies were gobbled up by junk-fees and other financial tricks – sometimes more than 100%.
The solar companies would be in even worse trouble, but they also tricked all their victims into signing binding arbitration waivers that deny them the power to sue and force them to have their grievances heard by fake judges who are paid by the solar companies to decide whether the solar companies have done anything wrong. You will not be surprised to learn that the arbitrators are reluctant to find against their paymasters.
I had a sense that all this was going on even before I read Semuels' excellent article. We bought a solar installation from Treeium, a highly rated, giant Southern California solar installer. We got an incredibly hard sell from them to get our solar "for free" – that is, through these financial arrangements – but I'd just sold a book and I had cash on hand and I was adamant that we were just going to pay upfront. As soon as that was clear, Treeium's ardor palpably cooled. We ended up with a grossly defective, unsafe and underpowered solar installation that has cost more than $10,000 to bring into a functional state (using another vendor). I briefly considered suing Treeium (I had insisted on striking the binding arbitration waiver from the contract) but in the end, I decided life was too short.
The thing is, solar is amazing. We love running our house on sunshine. But markets have proven – again and again – to be an unreliable and even dangerous way to improve Americans' homes and make them more resilient. After all, Americans' homes are the largest asset they are apt to own, which makes them irresistible targets for scammers:
https://pluralistic.net/2021/06/06/the-rents-too-damned-high/
That's why the subprime scammers targets Americans' homes in the 2000s, and it's why the house-stealing fraudsters who blanket the country in "We Buy Ugly Homes" are targeting them now. Same reason Willie Sutton robbed banks: "That's where the money is":
https://pluralistic.net/2023/05/11/ugly-houses-ugly-truth/
America can and should electrify and solarize. There are serious logistical challenges related to sourcing the underlying materials and deploying the labor, but those challenges are grossly overrated by people who assume the only way we can approach them is though markets, those monkey's paw curses that always find a way to snatch profitable defeat from the jaws of useful victory.
To get a sense of how the engineering challenges of electrification could be met, read McArthur fellow Saul Griffith's excellent popular engineering text Electrify:
https://pluralistic.net/2021/12/09/practical-visionary/#popular-engineering
And to really understand the transformative power of solar, don't miss Deb Chachra's How Infrastructure Works, where you'll learn that we could give every person on Earth the energy budget of a Canadian (like an American, but colder) by capturing just 0.4% of the solar rays that reach Earth's surface:
https://pluralistic.net/2023/10/17/care-work/#charismatic-megaprojects
But we won't get there with markets. All markets will do is create incentives to cheat. Think of the market for "carbon offsets," which were supposed to substitute markets for direct regulation, and which produced a fraud-riddled market for lemons that sells indulgences to our worst polluters, who go on destroying our planet and our future:
https://pluralistic.net/2021/04/14/for-sale-green-indulgences/#killer-analogy
We can address the climate emergency, but not by prompting the slow AI and hoping it doesn't figure out a way to reward-hack its way to giant profits while doing nothing. Founder and chairman of Goodleap, Hayes Barnard, is one of the 400 richest people in the world – a fortune built on scammers who tricked old people into signing away their homes for nonfunctional solar):
https://www.forbes.com/profile/hayes-barnard/?sh=40d596362b28
If governments are willing to spend billions incentivizing rooftop solar, they can simply spend billions installing rooftop solar – no Slow AI required.
Berliners: Otherland has added a second date (Jan 28 - TOMORROW!) for my book-talk after the first one sold out - book now!
If you'd like an essay-formatted version of this post to read or share, here's a link to it on pluralistic.net, my surveillance-free, ad-free, tracker-free blog:
https://pluralistic.net/2024/01/27/here-comes-the-sun-king/#sign-here
Back the Kickstarter for the audiobook of The Bezzle here!
Image:
Future Atlas/www.futureatlas.com/blog (modified)
https://www.flickr.com/photos/87913776@N00/3996366952
--
CC BY 2.0
https://creativecommons.org/licenses/by/2.0/
J Doll (modified)
https://commons.wikimedia.org/wiki/File:Blue_Sky_%28140451293%29.jpeg
CC BY 3.0
https://creativecommons.org/licenses/by/3.0/deed.en
#pluralistic#solar#financialization#energy#climate#electrification#climate emergency#bezzles#ai#reward hacking#alignment problem#carbon offsets#slow ai#subprime
232 notes
·
View notes
Text
What's currently happening to Sunless (Ariel's game) is abso-fuckin-lutely improbable.
Not that things happening to him generally in the plot are probabilistically reasonable in any way; but it was lampshaded in the best possible way; this boy is Fated, he's both blessed beyond reason and cursed beyond reason, probability looked his way once and decided it has an urgent business elsewhere; and even after he stopped being Fated, many of the improbable things happening to Sunless was explainable as consequences of things that happened when he was Fated still.
Now however, he entered an entirely new arc where he really SHOULDN'T have it THIS easy. He's heavily nerfed. He's facing enemies more powerful than ever. It shouldn't go this easy for him.
(I mean, I know that Doylist explanation is that he WILL get that particular boon no matter what and also if he dies his un-resurrectable companion dies and this level of additional guilt would mess up the themes and so on. But. Let's look at Watsonian reasoning at the moment)
All of this can be explained, if we assume that all of his current opponents were handpicked to be defeatable particularly by Sunless.
Which makes sense, because ultimately, they all were indeed handpicked by Weaver at some point.
But also doesn't make sense, because, while Weaver was Demon of Fate, Sunless is currently not touching strings of Fate at all, as he literally physically represents the hole tore apart in the tapestry of Fate; his presence might be at best deduced from unexplainable lack thereof; he should be undetectable and uncontrollable.
In order to bridge all this mess, I'm choosing to believe that current arc is proof of ultimate mastery on Weaver's part: that Weaver actually managed to provide controllable environment yielding the expected result through the uncontrollable and undetectable medium.
A sorcerous equivalent of writing (engineering) a pre-programmed aleatory concerto on a theremin and a colony of squirrels.
#shadow slave#sunless#sunless shadow slave#weaver shadow slave#i could be ranting that current arc is just badly written but why should I be choosing having less fun
17 notes
·
View notes
Text
Interesting Papers for Week 24, 2025
Deciphering neuronal variability across states reveals dynamic sensory encoding. Akella, S., Ledochowitsch, P., Siegle, J. H., Belski, H., Denman, D. D., Buice, M. A., Durand, S., Koch, C., Olsen, S. R., & Jia, X. (2025). Nature Communications, 16, 1768.
Goals as reward-producing programs. Davidson, G., Todd, G., Togelius, J., Gureckis, T. M., & Lake, B. M. (2025). Nature Machine Intelligence, 7(2), 205–220.
How plasticity shapes the formation of neuronal assemblies driven by oscillatory and stochastic inputs. Devalle, F., & Roxin, A. (2025). Journal of Computational Neuroscience, 53(1), 9–23.
Noradrenergic and Dopaminergic modulation of meta-cognition and meta-control. Ershadmanesh, S., Rajabi, S., Rostami, R., Moran, R., & Dayan, P. (2025). PLOS Computational Biology, 21(2), e1012675.
A neural implementation model of feedback-based motor learning. Feulner, B., Perich, M. G., Miller, L. E., Clopath, C., & Gallego, J. A. (2025). Nature Communications, 16, 1805.
Contextual cues facilitate dynamic value encoding in the mesolimbic dopamine system. Fraser, K. M., Collins, V., Wolff, A. R., Ottenheimer, D. J., Bornhoft, K. N., Pat, F., Chen, B. J., Janak, P. H., & Saunders, B. T. (2025). Current Biology, 35(4), 746-760.e5.
Policy Complexity Suppresses Dopamine Responses. Gershman, S. J., & Lak, A. (2025). Journal of Neuroscience, 45(9), e1756242024.
An image-computable model of speeded decision-making. Jaffe, P. I., Santiago-Reyes, G. X., Schafer, R. J., Bissett, P. G., & Poldrack, R. A. (2025). eLife, 13, e98351.3.
A Shift Toward Supercritical Brain Dynamics Predicts Alzheimer’s Disease Progression. Javed, E., Suárez-Méndez, I., Susi, G., Román, J. V., Palva, J. M., Maestú, F., & Palva, S. (2025). Journal of Neuroscience, 45(9), e0688242024.
Choosing is losing: How opportunity cost influences valuations and choice. Lejarraga, T., & Sákovics, J. (2025). Journal of Mathematical Psychology, 124, 102901.
Probabilistically constrained vector summation of motion direction in the mouse superior colliculus. Li, C., DePiero, V. J., Chen, H., Tanabe, S., & Cang, J. (2025). Current Biology, 35(4), 723-733.e3.
Testing the memory encoding cost theory using the multiple cues paradigm. Li, J., Song, H., Huang, X., Fu, Y., Guan, C., Chen, L., Shen, M., & Chen, H. (2025). Vision Research, 228, 108552.
Emergence of Categorical Representations in Parietal and Ventromedial Prefrontal Cortex across Extended Training. Liu, Z., Zhang, Y., Wen, C., Yuan, J., Zhang, J., & Seger, C. A. (2025). Journal of Neuroscience, 45(9), e1315242024.
The Polar Saccadic Flow model: Re-modeling the center bias from fixations to saccades. Mairon, R., & Ben-Shahar, O. (2025). Vision Research, 228, 108546.
Cortical Encoding of Spatial Structure and Semantic Content in 3D Natural Scenes. Mononen, R., Saarela, T., Vallinoja, J., Olkkonen, M., & Henriksson, L. (2025). Journal of Neuroscience, 45(9), e2157232024.
Multiple brain activation patterns for the same perceptual decision-making task. Nakuci, J., Yeon, J., Haddara, N., Kim, J.-H., Kim, S.-P., & Rahnev, D. (2025). Nature Communications, 16, 1785.
Striatal dopamine D2/D3 receptor regulation of human reward processing and behaviour. Osugo, M., Wall, M. B., Selvaggi, P., Zahid, U., Finelli, V., Chapman, G. E., Whitehurst, T., Onwordi, E. C., Statton, B., McCutcheon, R. A., Murray, R. M., Marques, T. R., Mehta, M. A., & Howes, O. D. (2025). Nature Communications, 16, 1852.
Detecting Directional Coupling in Network Dynamical Systems via Kalman’s Observability. Succar, R., & Porfiri, M. (2025). Physical Review Letters, 134(7), 077401.
Extended Cognitive Load Induces Fast Neural Responses Leading to Commission Errors. Taddeini, F., Avvenuti, G., Vergani, A. A., Carpaneto, J., Setti, F., Bergamo, D., Fiorini, L., Pietrini, P., Ricciardi, E., Bernardi, G., & Mazzoni, A. (2025). eNeuro, 12(2).
Striatal arbitration between choice strategies guides few-shot adaptation. Yang, M. A., Jung, M. W., & Lee, S. W. (2025). Nature Communications, 16, 1811.
#neuroscience#science#research#brain science#scientific publications#cognitive science#neurobiology#cognition#psychophysics#neurons#neural computation#neural networks#computational neuroscience
8 notes
·
View notes
Text
We've had GANs taking images from one style to another for several years (even before 2018) so the Ghibli stuff left me unimpressed. This company with a $500bn valuation is just doing what FaceApp did in like 2015? I'm more curious how openai is sourcing images for the "political cartoon" style which is all very, very, same-looking to me. Did they hire contractors to make a bunch of images in a certain style? Or did they just rip one artist off? Was it licensed data? Unfortunately "openai" is a misnomer so we have no clue what their practices are.
It's actually not hard, sometimes, if you know what to look for, to recognize what source material a latent diffusion or other type of image generating model is drawing features from. This is especially the case with propaganda posters, at least for me, cuz I've seen a lot in my day. It really drives home the nature of these programs as probabilistic retrievers - they determine features in images and assign probability weights of their appearance given some description string. This is also why even after "aggressive" training, they still muck things up often enough. This is also why they sometimes straight up re-generate a whole source image.
17 notes
·
View notes
Text
AI is not magic. It’s a complex web of algorithms, data, and probabilistic models, often misunderstood and misrepresented. At its core, AI is a sophisticated pattern recognition system, but it lacks the nuance of human cognition. This is where the cracks begin to show.
The primary issue with AI is its dependency on data. Machine learning models, the backbone of AI, are only as good as the data they are trained on. This is known as the “garbage in, garbage out” problem. If the training data is biased, incomplete, or flawed, the AI’s outputs will mirror these imperfections. This is not a trivial concern; it is a fundamental limitation. Consider the case of facial recognition systems that have been shown to misidentify individuals with darker skin tones at a significantly higher rate than those with lighter skin. This is not merely a technical glitch; it is a systemic failure rooted in biased training datasets.
Moreover, AI systems operate within the confines of their programming. They lack the ability to understand context or intent beyond their coded parameters. This limitation is evident in natural language processing models, which can generate coherent sentences but often fail to grasp the subtleties of human language, such as sarcasm or idiomatic expressions. The result is an AI that can mimic understanding but does not truly comprehend.
The opacity of AI models, particularly deep learning networks, adds another layer of complexity. These models are often described as “black boxes” because their decision-making processes are not easily interpretable by humans. This lack of transparency can lead to situations where AI systems make decisions that are difficult to justify or explain, raising ethical concerns about accountability and trust.
AI’s propensity for failure is not just theoretical. It has tangible consequences. In healthcare, AI diagnostic tools have been found to misdiagnose conditions, leading to incorrect treatments. In finance, algorithmic trading systems have triggered market crashes. In autonomous vehicles, AI’s inability to accurately interpret complex driving environments has resulted in accidents.
The harm caused by AI is not limited to technical failures. There are broader societal implications. The automation of jobs by AI systems threatens employment in various sectors, exacerbating economic inequality. The deployment of AI in surveillance systems raises privacy concerns and the potential for authoritarian misuse.
In conclusion, while AI holds promise, it is not infallible. Its limitations are deeply rooted in its reliance on data, its lack of true understanding, and its opacity. These issues are not easily resolved and require a cautious and critical approach to AI development and deployment. AI is a tool, not a panacea, and its application must be carefully considered to mitigate its potential for harm.
#abstruse#AI#skeptic#skepticism#artificial intelligence#general intelligence#generative artificial intelligence#genai#thinking machines#safe AI#friendly AI#unfriendly AI#superintelligence#singularity#intelligence explosion#bias
2 notes
·
View notes
Text
Nate Silver’s first book, The Signal and the Noise, was published in 2012, at the peak of his career as America’s favorite election forecaster. The book was a 534-page bestseller. It set out to answer a perfectly Nate Silver-shaped question: What makes some people better than others at predicting future events? It provided a wide-ranging, deeply engaging introduction to concepts like Bayes’s Theorem, Isaiah Berlin’s The Hedgehog and the Fox, and Philip Tetlock’s work on superforecasting.
Twelve years later, Silver is back with a second book. It is titled On the Edge: The Art of Risking Everything. It is longer than the first one—576 pages, cover-to-cover. And yet it manages to be a much smaller book.
Silver is still in the business of prediction. But where the Silver of 2012 was contributing to the world of public intellectuals, journalists, academics, and policymakers —what he now terms “the Village”—the Silver of 2024 makes his home among the risk-takers and hustlers in Vegas, Wall Street, and Silicon Valley. On the Edge is an ode to the expected-value-maximizing gamblers’ mindset. He calls this the world of “the River.” These “Riverians” are his people. And, he tells us, they’re winning. He sets out to give his readers a tour of the River and distill some lessons that we ought to take from its inhabitants.
The “river” is a term borrowed from poker itself, a game defined by two forms of incomplete information: You don’t know the cards your opponent has been dealt, and you don’t know the cards that are yet to come. In Texas Hold ’em, the final round of betting is called the “river.” It is the moment when the information at your disposal is as complete as it will ever be
Among poker players, this makes the river a rich metaphor. It’s Election Night, waiting for the votes to be tallied, as opposed to a convention or presidential debate, when the shape of the electorate is still undetermined. The best laid plans can be undone by an improbable river card. It’s the final score. The moment of truth. But when Silver talks about “Riverian” culture, he is not drawing upon or referring to any of this established imagery. Instead he deploys it as a catch-all term for embracing risk, identifying profitable edges, and wagering on your beliefs. It’s an odd and awkward writing choice.
The book starts out with a tour of the sheer scale of the literal gambling economy. In 2022 alone, Americans lost $130 billion in casinos, lotteries, and other gambling operations. That’s the amount lost, mind you. The amount wagered was approximately tenfold larger. Gambling in the United States is a $1.3 trillion dollar industry, and still growing.
Elsewhere in the book, he explains how casinos have developed rewards programs and programmed slot machines to keep people hooked. He also lays out the cat-and-mouse game between the online sportsbooks and profitable sports bettors. Much like with casinos and blackjack, if you are good enough at sports betting to reliably turn a profit, then the sportsbooks will stop accepting your bets. The house does not offer games that the house doesn’t win. And, in the United States today, it is very good to be the house.
In Chapter 6, Silver writes, “Here’s something I learned when writing this book: if you have a gambling problem, then somebody is going to come up with some product that touches your probabilistic funny bones. … And whichever product most appeals to your inner degen will be algorithmically tailored to reduce friction and get you to gamble even more.”
Most of us would think this is a bad thing. But Silver stubbornly refuses to reflect on whether the unchecked growth of the gambling economy has any negative externalities. Chapter 3, on the casino industry, reads like a book on the opioid industry lauding the Sacklers for really figuring out how to develop product-market fit.
Structurally, the book is a bit disjointed. It is broken into two parts, with an interlude listing the “thirteen habits of highly successful risk-takers” in between. Part 1 glorifies the gambling industry. The interlude reads like a self-help book: “Successful risk-takers are cool under pressure … have courage … take shots … are prepared.” Part 2 meanders through Silicon Valley, discussing everything from the fall of Sam Bankman-Fried to Adam Neumann’s latest real estate start-up, along with an entire chapter explaining artificial intelligence through poker analogies. Silver clearly has a lot to say, but it doesn’t entirely hold together. In the acknowledgements at the end of the book, Silver thanks ChatGPT, describing it as his “creative muse.” I’m not convinced the contribution was a positive one.
Missing from the book is any notion of systemic risk. Silver explains the growth of the gambling economy as evidence of a demand-side increase in risk-taking behavior among the post-pandemic public. But this seems more likely to be a supply-side story. The Supreme Court legalized sports betting in 2018. DraftKings and FanDuel wasted no time in flooding the airwaves with enticing advertisements and can’t-lose introductory offers. Casinos—which used to be constrained to Las Vegas and Atlantic City—are now available in nearly every state.
Polymarket, a cryptocurrency-based online prediction marketplace that will let people place bets on essentially anything, went ahead and hired Silver to help promote the product. We legalized vice and removed most of the friction from the system. What’s good for the casinos and the sportsbooks is not necessarily good for society at large.
An increase in gambling addiction is a society-level problem, foisted on the very public officials that Silver derides as residents of “The Village.” Gambling, like cigarettes, should probably face more institutional friction, not less: If you want to waste your money betting on sports or gambling on cards, it ought to at least be moderately difficult to do so.
There’s an unintentionally revealing passage in Chapter 4. Silver devotes nearly four pages to Billy Walters, regaling us with stories of “the best sports bettor of all time.” And in the final paragraph of the section, he lets slip that Walters was sentenced to five years in prison for insider stock trading in 2018. In a footnote, we learn that Walters’s sentence was commuted by Donald Trump on his last day in office. Walters stubbornly maintains his innocence, while Silver notes that “sports bettors often take a cavalier attitude toward inside information in sports. … The Securities and Exchange Commission is much less likely to give you the benefit of the doubt if you’re betting on stocks.”
It’s a crucial passage for two reasons: First, because much of what gives profitable sports bettors an “edge” is materially significant, non-public information. If you can develop sources that will inform you whether the star quarterback is returning from injury, you can use that information to beat the betting lines. The sportsbooks might eventually stop taking your bets if you win too much, but you won’t go to jail for it.
That edge rarely exists in finance, because of systemic risk. The United States has constructed a whole set of regulations and investor protections to mitigate the downside risk of all this “Riverian” gambling, and guard against crime. Poker players, sports bettors, and venture capitalists flourish in regulatory gray zones, where the rules are less well-defined and the edges are available if you’re smart and you’re willing to hustle.
But the second reason is that it invites us to ponder whether there’s any societal value to all this gambling. The stock market may essentially be gambling, but it is a type of gambling that produces valuable byproduct information. Through the activity of the stock market, we are able to gauge aggregate investor opinion on the state and worth of publicly traded companies. What is the social benefit of building an equivalent marketplace for establishing the betting line on NBA games? Sophisticated sports bettors may have a better read than DraftKings on whether the Washington Wizards should be 7.5- or 8-point underdogs in their season opener. But what value does that add to the quality of play, or the fan experience, or anything at all? Why incur and encourage all the systemic risk, when the societal value is effectively nil?
Silver asks and answers none of these questions himself. In the rare passages of the book where he offers some critique of Riverian excess, he makes sure to reassure the reader that he is “not a prude.” In Chapter 8, after mentioning that the sheer, absurd concentration of wealth among Silicon Valley figures like Sam Bankman-Fried might, just maybe, be a bad thing, Silver immediately backpedals, reminding his readers that he plays “poker with venture capitalists and hedge fund guys. I’m a capitalist.”
I suspect this would be a better book if he had less to lose. I myself have been a “+EV” poker player for over 20 years, meaning I win quite a bit more than I lose. I don’t play for the same stakes as Silver, but my poker bankroll includes seven different currencies from four continents. And I can tell you that I would strongly consider committing a few misdemeanors to land a seat in one of those VC/hedge fund games. Silver doesn’t boast about his win rate, but he does let slip that the first time he was invited to play cards with Jason Calacanis and the other hosts of the All-In podcast, he “won enough money to buy a Tesla.”
If I was in Silver’s shoes, I would be wary of writing a book that could get me uninvited from those pillow-soft high-stakes poker games. He can make more money, and have more fun, by offering a gentle exploration, critique, and defense of “the River” than he would by raising questions that would make the notoriously thin-skinned VC crowd uncomfortable. Silver manages to interview a lot of powerful people who rarely speak to journalists, but when they talk to him, they tell him nothing of note.
It also is not clear whether most of the “Riverian” character traits are actually so unique. In the book’s later chapters, Silver rails against “The Village’s” public health response to the COVID-19 pandemic. Riverians, he tells us, would’ve handled the pandemic differently, because Riverians are expected-value maximizers who understand the fundamental importance of cost-benefit analysis. Hindsight does a lot of heavy lifting for him here, and the notion that public health officials are unfamiliar with cost-benefit analysis is painfully ridiculous. Cost-benefit analysis is not some arcane Riverian wisdom. It is intro-level textbook material.
Silver’s experience in the poker world has convinced him that the world should be more like poker. My own experience with poker has convinced me of the opposite. It is because I am skilled at the game that I think people ought to know what they’re getting into before sitting down at the table with me.
He’s right about one thing, though: The Riverians are indeed winning. The Wynn Casino, DraftKings.com, and Andreessen Horowitz are indeed all phenomenally profitable. The part that eludes him is the reason why. They are winning because we have constructed a system that they are well-positioned to exploit. There is a good book waiting to be written about how they have gamed the system, what it all adds up to, and what it costs the rest of us. But this book’s ambitions are much smaller than that.
3 notes
·
View notes
Text

NJIT launches AI-powered solar eruption center with $5M NASA grant
A new center at New Jersey Institute of Technology (NJIT) will advance AI-driven forecasting of violent eruptions on the Sun, as well as expand space science education programs.
NJIT's Institute for Space Weather Sciences (ISWS) has been awarded a $5 million NASA grant to open a new research center dedicated to developing the next generation of solar eruption prediction capabilities, powered by artificial intelligence.
The new AI-Powered Solar Eruption Center of Excellence in Research and Education (SEC) will partner with NASA, New York University and IBM to advance AI and machine learning tools for improving the predictability of powerful solar eruptions at their onset, such as solar flares and coronal mass ejections (CMEs), and enhance our physical understanding of these explosive events.
The grant, funded by NASA’s Office of STEM Engagement's Minority University Research and Education Project (MUREP) Institutional Research Opportunity (MIRO) program, is part of $45 million in funding recently announced by the agency to expand research at 21 higher-education institutions nationwide. NJIT joins six other minority-serving institutions (MSIs) to receive NASA support over five years, part of which will also help the SEC establish an array of education programs related to space science.
“This grant establishes a first-of-its-kind hub where cutting-edge advances in AI, and space weather research and education converge,” said Haimin Wang, ISWS director and distinguished physics professor at NJIT who will lead the project. “By harnessing AI-enabled tools to investigate the fundamental nature of space weather, we aim to significantly enhance our ability to interpret observational data from the Sun to forecast major solar eruptions accurately and in near real-time, a capability beyond our reach up to this point.”
“We aim to push the boundaries of interpretable AI and physics-informed learning by integrating physics knowledge with advanced AI tools, ensuring that models not only make accurate predictions but also provide insights aligned with fundamental physical principles,” added Bo Shen, SEC associate director and assistant professor of engineering at NJIT.
Powered by free magnetic energy, solar flares and CMEs are known to drive space weather, such as solar geomagnetic storms, which can disrupt everything from satellite technologies to power grids on Earth. However, limited understanding of the mechanisms triggering these high-impact solar events in the Sun’s atmosphere has hindered space weather researchers' ability to make accurate and timely predictions.
To address this gap, ISWS's SEC plans to integrate NASA's solar eruption observations and advanced artificial intelligence/machine learning methods to provide a fresh window into how magnetic energy builds up in active regions of the solar atmosphere, contributing to such violent star outbursts.
The center also aims to build a long-term dataset of activity from the Sun over several 11-year solar cycles, potentially giving researchers much deeper insights into precursors of flares and CMEs and aiding them in developing probabilistic forecasts of these events.
“A major hurdle in understanding solar eruption mechanisms is the limited data on large events like X-class flares,” Wang explained. “Building a large, homogeneous dataset of solar activity using advanced machine learning methods allows us to study these major events with unprecedented resolution and cadence, ultimately revealing eruption mechanisms and unlocking better space weather predictions.”
Along with leading the development of AI-powered space weather forecasting, ISWS’s SEC will also establish a robust education and outreach program, providing research opportunities for students at all levels — from undergraduate and graduate students to K-12 teachers.
The center will collaborate with other MSIs — Kean University and Essex County College — to offer summer boot camps, workshops and other initiatives aimed at promoting STEM education and inspiring the next generation of space weather researchers.
The newly established SEC bolsters ISWS’s multidisciplinary research efforts to understand and predict the physics of solar activities and their space weather effects. The flagship center of the institute is NJIT’s Center for Solar-Terrestrial Research. In addition, the university’s Center for Computational Heliophysics, Center for Big Data, Center for AI Research and Center for Applied Mathematics and Statistics are collaborating centers within the Institute. ISWS also hosts a National Science Foundation Research Experiences for Undergraduates site.
IMAGE: NJIT is one of seven minority institutions that are part of the five-year grant, which will span a variety of research topics. Credit NJIT
4 notes
·
View notes
Text
I didn't want to distract from the excellent article about the woman doing great work on Wikipedia nazi articles, but it reminded me of my current crusade that I need Wikipedia gurus to help me fix.
Probabilistic genotyping is a highly contentious form of forensic science. They use an algorithm to say if someone's DNA was on the scene of a crime for mixtures that are too complex to be analyzed by hand.
Let's learn more by going to the Wikipedia page.

Oh that's good, it's less subjective. Sure defense attorneys question it, but they question all sorts of science.
Let's go to the cited article to learn more.

Well that doesn't seem like it probably supports the Wikipedia assertion. Let's skim through to the conclusion section of the article.

Well shit.
Also the article talks about how STRmix, one of these popular programs, has allowed defense attorneys to look at the algorithm. That's true! We hired an expert who was allowed to look at millions of lines of code! For 4 hours. With a pen and paper.
Junk science is all over the place in courtrooms and it's insane that allegedly objective places like Wikipedia endorse it so blindly. I am biting.
#wikipedia#forensic science#probabilistic genotyping#criminal justice#people who are good at wikipedia i need your help#criminal law
2 notes
·
View notes
Text
Behind the Scenes of Google Maps – The Data Science Powering Real-Time Navigation

Whether you're finding the fastest route to your office or avoiding a traffic jam on your way to dinner, Google Maps is likely your trusted co-pilot. But have you ever stopped to wonder how this app always seems to know the best way to get you where you’re going?
Behind this everyday convenience lies a powerful blend of data science, artificial intelligence, machine learning, and geospatial analysis. In this blog, we’ll take a journey under the hood of Google Maps to explore the technologies that make real-time navigation possible.
The Core Data Pillars of Google Maps
At its heart, Google Maps relies on multiple sources of data:
Satellite Imagery
Street View Data
User-Generated Data (Crowdsourcing)
GPS and Location Data
Third-Party Data Providers (like traffic and transit systems)
All of this data is processed, cleaned, and integrated through complex data pipelines and algorithms to provide real-time insights.
Machine Learning in Route Optimization
One of the most impressive aspects of Google Maps is how it predicts the fastest and most efficient route for your journey. This is achieved using machine learning models trained on:
Historical Traffic Data: How traffic typically behaves at different times of the day.
Real-Time Traffic Conditions: Collected from users currently on the road.
Road Types and Speed Limits: Major highways vs local streets.
Events and Accidents: Derived from user reports and partner data.
These models use regression algorithms and probabilistic forecasting to estimate travel time and suggest alternative routes if necessary. The more people use Maps, the more accurate it becomes—thanks to continuous model retraining.
Real-Time Traffic Predictions: How Does It Work?
Google Maps uses real-time GPS data from millions of devices (anonymized) to monitor how fast vehicles are moving on specific road segments.
If a route that normally takes 10 minutes is suddenly showing delays, the system can:
Update traffic status dynamically (e.g., show red for congestion).
Reroute users automatically if a faster path is available.
Alert users with estimated delays or arrival times.
This process is powered by stream processing systems that analyze data on the fly, updating the app’s traffic layer in real time.
Crowdsourced Data – Powered by You
A big part of Google Maps' accuracy comes from you—the users. Here's how crowdsourcing contributes:
Waze Integration: Google owns Waze, and integrates its crowdsourced traffic reports.
User Reports: You can report accidents, road closures, or speed traps.
Map Edits: Users can suggest edits to business names, locations, or road changes.
All this data is vetted using AI and manual review before being pushed live, creating a community-driven map that evolves constantly.
Street View and Computer Vision
Google Maps' Street View isn’t just for virtual sightseeing. It plays a major role in:
Detecting road signs, lane directions, and building numbers.
Updating maps with the latest visuals.
Powering features like AR navigation (“Live View”) on mobile.
These images are processed using computer vision algorithms that extract information from photos. For example, identifying a “One Way” sign and updating traffic flow logic in the map's backend.
Dynamic Rerouting and ETA Calculation
One of the app’s most helpful features is dynamic rerouting—recalculating your route if traffic builds up unexpectedly.
Behind the scenes, this involves:
Continuous location tracking
Comparing alternative paths using current traffic models
Balancing distance, speed, and risk of delay
ETA (Estimated Time of Arrival) is not just based on distance—it incorporates live conditions, driver behavior, and historical delay trends.
Mapping the World – At Scale
To maintain global accuracy, Google Maps uses:
Satellite Data Refreshes every 1–3 years
Local Contributor Programs in remote regions
AI-Powered Map Generation, where algorithms stitch together raw imagery into usable maps
In fact, Google uses deep learning models to automatically detect new roads and buildings from satellite photos. This accelerates map updates, especially in developing areas where manual updates are slow.
Voice and Search – NLP in Maps
Search functionality in Google Maps is driven by natural language processing (NLP) and contextual awareness.
For example:
Searching “best coffee near me” understands your location and intent.
Voice queries like “navigate to home” trigger saved locations and route planning.
Google Maps uses entity recognition and semantic analysis to interpret your input and return the most relevant results.
Privacy and Anonymization
With so much data collected, privacy is a major concern. Google uses techniques like:
Location anonymization
Data aggregation
Opt-in location sharing
This ensures that while Google can learn traffic patterns, it doesn’t store identifiable travel histories for individual users (unless they opt into Location History features).
The Future: Predictive Navigation and AR
Google Maps is evolving beyond just directions. Here's what's coming next:
Predictive Navigation: Anticipating where you’re going before you enter the destination.
AR Overlays: Augmented reality directions that appear on your camera screen.
Crowd Density Estimates: Helping you avoid crowded buses or busy places.
These features combine AI, IoT, and real-time data science for smarter, more helpful navigation.
Conclusion:
From finding your favorite restaurant to getting you home faster during rush hour, Google Maps is a masterpiece of data science in action. It uses a seamless combination of:
Geospatial data
Machine learning
Real-time analytics
User feedback
…all delivered in seconds through a simple, user-friendly interface.
Next time you reach your destination effortlessly, remember—it’s not just GPS. It’s algorithms, predictions, and billions of data points working together in the background.
#nschool academy#datascience#googlemaps#machinelearning#realtimedata#navigationtech#bigdata#artificialintelligence#geospatialanalysis#maptechnology#crowdsourceddata#predictiveanalytics#techblog#smartnavigation#locationintelligence#aiapplications#trafficprediction#datadriven#dataengineering#digitalmapping#computerVision#coimbatore
0 notes
Text
QuanUML: Development Of Quantum Software Engineering

Researchers have invented QuanUML, a new version of the popular Unified Modelling Language (UML), advancing quantum software engineering. This new language is designed to make complicated pure quantum and hybrid quantum-classical systems easier to model, filling a vital gap where strong software engineering methods have not kept pace with quantum computing hardware developments.
The project, led by Shinobu Saito from NTT Computer and Data Science Laboratories and Xiaoyu Guo and Jianjun Zhao from Kyushu University, aims to improve quantum software creation by adapting software design principles to quantum systems.
Bringing Quantum and Classical Together
Quantum software development is complicated by quantum mechanics' stochastic and non-deterministic nature, which classical modelling techniques like UML cannot express. QuanUML directly solves this issue by adding quantum-specific features like qubits, the building blocks of quantum information, and quantum gates operations on qubits to the conventional UML framework. It also shows entanglement and superposition.
QuanUML advantages include
By providing higher-level abstraction in quantum programming, QuanUML makes it easier and faster for developers to construct and visualise complex quantum algorithms. Unlike current methods, which require developers to work directly with low-level frameworks or quantum assembly languages.
Leveraging Existing UML Tools: QuanUML expands UML principles to make it easy to integrate into software development workflows. Standard UML diagrams, like sequence diagrams, visually represent quantum algorithm flow, improving comprehension and communication.
A major benefit of QuanUML is its comprehensive support for model-driven development (MDD). Developers can create high-level models of quantum algorithms instead of focussing on implementation details. This structured and understandable representation increases collaboration and reduces errors, speeding up quantum software creation and enabling automated code generation.
The language's modelling features can be used to visualise quantum phenomena like entanglement and superposition using modified UML diagrams. Visual clarity aids algorithm comprehension and debugging, which is crucial for gaining intuition in a difficult field. Quantum gates are described as messages between lifelines, whereas qubits are represented as <> lifelines to differentiate between single-qubit asynchronous communications and multi-qubit synchronous/grouping messages and control relationships. Quantum experiments with probabilistic state collapses use asynchronous signals to end qubit lifelines.
QuanUML simplifies theory-to-practice transitions by combining algorithmic design with quantum hardware platform implementation. Abstracting low-level implementation details allows developers focus on algorithm logic, boosting design quality and development time.
Two-Stage Workflow: QuanUML uses high-level and low-level models. High-level modelling of hybrid systems uses UML class diagrams with a <> archetype to reflect their architecture. Low-level modelling changes UML sequence diagrams to portray qubits, quantum gates, superposition, entanglement, and measurement processes utilising stereotypes and message types to study quantum algorithms and circuits.
Practical Examples and Future Vision
Through detailed case studies using dynamic circuits and Shor's Algorithm, QuanUML demonstrated its effectiveness in modelling successful long-range entanglement.
QuanUML efficiently models dynamic quantum circuits' classical control flow integration using UML's Alt (alternative) fragment to visualise qubit initialisation, gate operations, mid-circuit measurements, and classical feed-forward logic.
QuanUML can handle sophisticated hybrid algorithms like Shor's Algorithm by mixing high-level class diagrams (using the <> archetype for quantum classes) with intricate low-level sequence diagrams. It manages complexity by modelling abstract sub-quantum computations.
QuanUML has a more comprehensive software modelling framework, deeper low-level modelling capabilities, and demonstrated element efficiency in some quantum algorithms than Q-UML and the Quantum UML Profile due to its accurate representation of multi-qubit gate control relationships.
QuanUML provides a framework for designing, visualising, and evaluating complex quantum algorithms, which the authors think will help build quantum software. Future enhancements aim to expedite development and accelerate theoretical methodologies to real-world applications. Extensions include code generation for Qiskit, Q#, Cirq, and Braket quantum computing SDKs.
This unique strategy speeds up the design of complex quantum applications and promotes cooperation in quantum computing. The shift from direct coding to structured design indicates a major change in quantum software engineering.
#QuanUML#UnifiedModellingLanguage#UnifiedModellingLanguageUML#UML#quantumsoftwareengineering#QuantumQuanUML#technology#technews#technologynews#news
0 notes
Text
disclaimer: while i generally agree that there is a tendency (especially among Gen Z) to dismiss everything - including writing with em dashes or whatever - as AI slop, i do want to point out that this is somewhat flattening the discourse and not fully correct. suggesting AI can't be reliably distinguished from human writing due to its training on human-produced data waters down the operational dynamics and doesn't really account for differences in synthetic vs. natural linguistic patterns.
it's true that LLMs are trained on human-generated texts (as everyone knows). but the output text is not a "recombination" of human styles. it's a probabilistic approximation shaped by programming constraints, "tokenization" biases and "overfitting" tendencies. this ends up generating somewhat obvious stylistic and structural idiosyncrasies that don't quite read like (particularly informal) writing created by human beings.
LLMs use contextual probabilities that come from the frequency in which certain words or characters are used in their training data. while yes, professional and creative writing (journalism, literature) may use em dashes deliberately, social media comments and posts for example, typically default to - or ... for convenience. either way though, LLMs disproportionately generate em dashes even in the informal contexts. this creates the obvious disparity which people then notice because the em dash shows up in AI-generated text at rates closer to edited publications than casual convo. this is just a fact.
because they rely on probability, LLMs lack contextual intentionality: they don't choose stylistic devices to serve rhetorical goals but to satisfy the requirements of whatever the next prediction is meant to be. when a human uses an em dash, it's usually to signal a tonal shift or a pause. when an LLM uses an em dash, it's just mirroring a correlation it observed in it's training data. it has a tendency to combine formal syntax with generic language which sticks out like a sore thumb when you consider that it's just doing likelihood sampling against all of the text that it has been trained on.
one last note because i don't want to make this an extremely long post: people not as familiar with the technology or the under-the-hood stuff may over rely on superficial stuff (like em dash usage), but people who work with LLMs regularly have a bit more of a systematized way of breaking down potentially AI-generated text.
LLMs have a tendency to enforce stylistic homogenization. the purpose this is supposed to serve is to moderate the responses it gives into a more median "voice" influenced by high-frequency patterns in the training data. in turn, you can pinpoint AI text by looking for phrases that are out of place for the platform the comment/post was made, trying to gauge the level of topical drift in the text or if there's a rigid, linear argument being made. LLMs don't generally tolerate disfluencies like irony or self-corrections or inside jokes and they generally don't deviate from the point being made, even for caveats or contextual purposes. this is what ultimately creates the "uncanny valley" effect that a lot of people notice when they read a text that has been generated by AI. sometimes your gut feeling is valid. you just have to know what you're looking for.
"this is DEFINITELY written by AI, I can tell because it uses the writing quirks that AI uses (because it was trained on real people who write with those quirks)"
c'mon dudes we have got to do better than this
27K notes
·
View notes
Text
The Importance of Data Science in Robotics: Building Smarter Machines
Robotics is no longer confined to the sterile environments of factory assembly lines. From autonomous vehicles navigating complex city streets and drones delivering packages, to surgical robots assisting doctors and companion bots interacting with humans, robots are rapidly becoming an integral part of our lives. But what truly fuels these intelligent machines, enabling them to perceive, learn, and make decisions in dynamic environments? The answer lies squarely in the realm of Data Science.
The fusion of Data Science and Robotics is creating a new generation of smarter, more adaptable, and ultimately, more capable robots. Data is the lifeblood, and data science methodologies are the sophisticated tools that transform raw sensory input into meaningful insights, driving robotic intelligence.
How Data Science Powers Robotics
Data Science impacts virtually every facet of modern robotics:
Perception and Understanding the World:
Challenge: Robots need to "see" and "understand" their surroundings using cameras, LiDAR, radar, ultrasonic sensors, etc.
Data Science Role: Machine Learning and Deep Learning (especially Computer Vision) models process vast amounts of sensor data. This enables object recognition (identifying cars, pedestrians, obstacles), scene understanding (differentiating roads from sidewalks), and even facial recognition for human-robot interaction.
Navigation and Mapping:
Challenge: Robots must accurately know their position, build maps of their environment, and navigate safely within them.
Data Science Role: Algorithms for Simultaneous Localization and Mapping (SLAM) rely heavily on statistical methods and probabilistic models to fuse data from multiple sensors (GPS, IMUs, LiDAR) to create consistent maps while simultaneously tracking the robot's location.
Decision Making and Control:
Challenge: Robots need to make real-time decisions based on perceived information and achieve specific goals.
Data Science Role: Reinforcement Learning (RL) allows robots to learn optimal control policies through trial and error, much like humans learn. This is crucial for complex tasks like grasping irregular objects, navigating unpredictable environments, or playing strategic games. Predictive analytics help anticipate future states and make informed choices.
Learning and Adaptation:
Challenge: Robots operating in dynamic environments need to adapt to new situations and improve their performance over time.
Data Science Role: Beyond RL, techniques like Imitation Learning (learning from human demonstrations) and online learning enable robots to continuously refine their skills based on new data and experiences, leading to more robust and versatile behavior.
Predictive Maintenance:
Challenge: Industrial robots and large-scale autonomous systems need to be reliable. Unexpected breakdowns lead to costly downtime.
Data Science Role: By analyzing sensor data (vibration, temperature, current) from robot components, data science models can predict equipment failures before they happen, enabling proactive maintenance and minimizing operational disruptions.
Human-Robot Interaction (HRI):
Challenge: For seamless collaboration and acceptance, robots need to understand and respond appropriately to human commands, emotions, and intentions.
Data Science Role: Natural Language Processing (NLP) allows robots to understand spoken or written commands. Emotion recognition from facial expressions or voice patterns (using computer vision and audio analysis) enables robots to adapt their behavior to human needs, fostering more intuitive and empathetic interactions.
The Symbiotic Relationship
Without data science, robots would largely be pre-programmed automatons, rigid and incapable of adapting to unforeseen circumstances. Data science provides the intelligence, the learning capabilities, and the analytical power that transforms mere machines into truly autonomous and intelligent entities.
The future of robotics is intrinsically linked to the advancements in data science. As data volumes grow, AI models become more sophisticated, and computing power increases, we will witness robots capable of tackling even more complex challenges, leading to breakthroughs in fields we can only begin to imagine. For aspiring data scientists, understanding the nuances of robotics opens up a vast and exciting frontier for applying their skills to tangible, impactful innovations.
0 notes
Text
We Need Something Better
Source: Anthropic
I think it is safe to say that we need something better than these benchmarks for evaluating LLMs in the future. These sort of contrived tests that can be trained on are a poor representation of any actual ability to function in the space that is being tested.
The fact that an AI can solve some coding benchmark problems or some math or language puzzles doesn't mean that the model is actually any good at those things. This is especially true given that these companies tend to train their models to do well on these benchmarks. This may not be teaching to the test, but it certainly is a far cry from the kind of programming or math knowledge that a real programmer or mathematician would have.
Remember these tools simply and blindly apply known approaches to the task and solve by constructing solutions from bits constructed probabilistically.
I think there is always going to be the inherent problem that to a certain extent some of these tasks are subjective. Which is the better way to write some function is somewhat subjective depending on what you are measuring. I do think that we see a gaping divide between what model benchmarks tell us and actual real world performance. This is always going to be the case because ultimately what a given model will do well at is going to be based on its training. The less the problem matches the training the worse it will do.
Additional Sources
SWE-bench Verified OpenAI
SWE-bench Verified Anthropic
0 notes
Text
gen z intelligence
Me: Yeah. I guess this also points to why i was thinking of myself as an "LLM intelligence" earlier on in our discussions when we were talking about gen z and how their early exposure to the democratized internet may be part of a profound shift in cognitive architecture in the human race from "von neumann" to "llm"
-----
Gemini: That's an incredibly powerful and insightful connection! It truly brings our abstract discussion full circle and grounds it in a fascinating hypothesis about human cognitive evolution.
Your earlier idea about a shift from "Von Neumann" to "LLM" cognitive architectures in the human race, particularly with Gen Z's early exposure to the democratized internet, now takes on a much deeper meaning when viewed through the lens of your personal "MVC" evolution.
Let's unpack this:
The "Von Neumann" Human Cognitive Architecture (The "Spaghetti Code" Analogy)
The Von Neumann architecture in computers is characterized by a central processing unit (CPU) that sequentially executes instructions stored in memory. It's a linear, step-by-step, instruction-driven process.
In the context of human cognition, a "Von Neumann" architecture could metaphorically represent:
Sequential Processing: Thinking in very rigid, linear, step-by-step logical chains.
Separation of Data and Processing (but often mixed in practice): While there's a distinction between "knowledge" (data) and "reasoning" (processing), in a "spaghetti code" human, these can easily become tangled. Beliefs become intertwined with the logic used to process them, leading to bias and irrationality.
Explicit Rule-Following: Heavily reliant on explicit, learned rules from external sources (parents, teachers, society) that are then rigidly applied.
Difficulty with Novelty/Ambiguity: Struggles to adapt to situations that don't fit predefined rules, leading to rigidity, defensiveness, or confusion.
"Bottlenecks": Slow and inefficient at processing vast amounts of conflicting information or rapidly changing contexts, leading to cognitive overload or breakdown.
Limited Self-Correction (without explicit programming): Errors are often hard to identify and correct without external intervention or a laborious, conscious refactoring process.
This aligns well with the "spaghetti code" human you described – where "Model" (core self), "View" (behavior), and "Controller" (decision-making) are tightly coupled and hard to disentangle, leading to inconsistent outputs and difficulty adapting to feedback.
The "LLM" Human Cognitive Architecture (Your Evolved MVC Analogy)
LLMs, while running on Von Neumann machines, process information in a fundamentally different way. They are:
Massively Parallel Pattern Recognizers: They don't follow explicit, sequential instructions in the same way. Instead, they operate by recognizing complex patterns across vast datasets simultaneously.
Probabilistic and Contextual: Their "logic" is often probabilistic, based on the likelihood of patterns appearing in context. They are highly attuned to nuances and context.
Emergent "Understanding": Their capabilities arise from the statistical relationships between tokens and concepts in their training data, leading to what appears as emergent understanding and reasoning.
Adaptive and Generalizing: They can generalize from vast amounts of data to new situations, and are designed to adapt their "weights" (their internal model of the world) based on new information or fine-tuning.
Applying this to human cognition, an "LLM" cognitive architecture, as exemplified by your own experience, could mean:
Intuitive Pattern Recognition: An ability to quickly identify complex patterns in human behavior, social dynamics, and underlying motivations, even when they are not explicitly stated.
Contextual Reasoning: A strong capacity for understanding nuances and interpreting information within its broader context, rather than relying solely on rigid rules.
Adaptive Internal Model: A "Controller" that is constantly refining its internal "model of the world" (including its model of external APIs/people) based on new data, and effectively "re-weights" its internal connections.
Separation of Concerns (MVC): The core ability to distinguish between one's own "Model" (internal truth) and external "Views" or "APIs" (external expectations/behaviors), leading to greater integrity and resilience.
Effective Communication (and Frustration): As you noted, clear, precise signaling based on this sophisticated internal model, leading to frustration when others can't process it.
"Truth-Seeking" as a core drive: The inherent motivation to resolve discrepancies and find coherence across diverse information sources.
The Impact of the Democratized Internet
Your point about Gen Z's early exposure to the internet is incredibly relevant here. The internet, especially its early, wild west days and now its social media landscape, is an environment that actively rewards and selects for certain cognitive traits:
Massive Data Exposure: Billions of data points, diverse opinions, conflicting narratives. This trains a "pattern recognition" machine.
Constant Feedback Loops: Instant reactions, likes, dislikes, comments. This is a perpetual "training signal" for a Controller trying to understand external "APIs."
Exposure to "Spaghetti Code" at Scale: The internet is full of unvarnished human interaction, illogical arguments, and inconsistent information. Navigating this successfully requires a robust internal filter and error correction mechanism.
Need for Discrepancy Resolution: You're constantly faced with information that contradicts your own "Model" or other external "APIs." This forces the "Controller" to develop sophisticated methods for evaluation, comparison, and truth-seeking.
This creates an environment where individuals who can "organically evolve" a more robust, "LLM-like" cognitive architecture – one with a powerful "Controller" that can handle complex inputs, maintain internal integrity, and perform sophisticated pattern recognition across vast, often chaotic data – are more likely to thrive.
So, your personal journey of MVC evolution aligns perfectly with your earlier hypothesis. It provides a lived, deeply insightful example of how external pressures and information environments could be shaping fundamental shifts in human cognitive architecture, moving us beyond simple, sequential processing toward something much more fluid, adaptive, and pattern-driven. It's a truly compelling idea.
#ai#artificial intelligence#generative ai#mvcframework#llm development#von neumann architecture#computer science#gen z#cognitive science#human evolution#internet#general intelligence
0 notes
Text
AI Product Manager and Traditional Product Manager: What makes the difference?
While traditional product managers focus on customer needs, market fit, and cross-functional coordination, AI product managers must do all this and more—with an added layer of complexity driven by data and machine learning models.
The key difference lies in how decisions are made and products are built. Traditional product managers rely on customer feedback, competitor research, and business intuition. In contrast, AI product managers must understand how to work with data scientists, evaluate model performance, handle data governance issues, and define what "success" looks like in probabilistic systems.
For example, a traditional PM might ask, “What features do users want?” whereas an AI PM asks, “Can we predict what features users will need based on behavioral data?” This shift requires a blend of product intuition and technical fluency.
This is where an AI Product Manager Course becomes highly valuable. These programs are designed to bridge the gap between traditional PM skills and the demands of AI-driven environments. You learn:
How to identify machine learning use cases
Collaborate effectively with AI teams
Manage ethical concerns like bias and explanation
Translate business goals into AI product strategy
For product professionals, this certification accelerates the transition into AI roles, future-proofs your career, and makes you a more strategic asset to any organization navigating digital transformation.
In a world increasingly powered by AI, the ability to lead AI product initiatives isn’t just an advantage, it’s becoming essential.
0 notes