#automated machine
Explore tagged Tumblr posts
Text

Whispering secret data.
#lab#machine#automation#robotics#cyberpunk#retro#scifi#stuck#laboratory#farm#android#cyborg#data#secret#whisper#illustration#drawing#digitalartwork#digitaldrawing#digitalart#digitalillustration#90s#cables#machinelearning#connection#ring#runner#net#flesh
5K notes
¡
View notes
Text
Belobog has a graveyard for broken down machines, but where do cyborgs go?
#honkai star rail#hsr#boothill#my art#chainslobber#âOh that's gore. That's gore of my beloved.â#we're still married I just put him in a car wash automated machine he's all better#cw: body horror#cw: gore
249 notes
¡
View notes
Text
Rahhhh finished this thing thatâs been in wip for a few months
One of 2 of my flightrising dragons who are automatons (they are identical except their magic/flight)
#myart#flight rising#flight rising fanart#flight rising ridgeback#dragon#fr fanart#fr ridgeback#automation#machine#robot
680 notes
¡
View notes
Text

Build-A-Bear vending machine, JFK Airport (Terminal 4)
#atm#automated teddy machine#build a bear vending machine#build a bear#vending machine#kidcore#2000s childhood#2000s nostalgia#nostalgiacore#plushies#plush#stuffed animals#stuffies#plush toy#soft toy#plushblr#bab#sfw agere#sfw agedre#agere#agedre
2K notes
¡
View notes
Text
Task Force 141âs go to Tesco Meal Deals
the tescos outside my uni is never fucking stocked i want the sandwich on ghosts so bad but that shit is always gone
John Price

- gets water as his drink (criminal)
- gets the mixed nuts as his snack (criminal)
- heâs smart tho he always gets the boujie water because the meal deal price is set
- the wrap is valid
- usually will also grab another snack bc this isnât enough maybe like a bag of dried mango or some shit (old)
Kyle âGazâ Garrick

- my headcanon continues to live
- the healthy energy shit tastes like ass but he pretends its good
- does not usually shop at tesco, heâs a waitrose boy
- sometimes indulges in the odd crunchie bar but rarely
John âSoapâ Mactavish

- horrifying (i love pepperami sm everyone disses me for it bc i pull that shit out in lesson and it stanks)
- protein to the max ig
- the whole meal fucking stinks
- uses gazâs club card because heâs too lazy to get his own
Simon âGhostâ Riley

- absolutely classic
- everytime he gets it mentions how cost of living prices have made them more expensive
- grenade bars are disgusting but he loves them for some reason (masochism imo)
- the sandwich is the best one they have bc the bread is always so moist its so fucking good
- the monster bc yall saw the ghost monster can we all know
#i need to sleep but this was important#i am very confident this is canon ty#at the self checkout gaz checks himself out in the screen thing#see what i did there#whenever soap finishes paying at tesco defo says the automated âthanks for shopping at tescoâ in unison with the machine#thinks its funny#its not#(me)#ghost is such a tesco trooper he still has the physical card club cards#soap and ghost went to sainsburys once and ghost wouldnt stop grumbling about how the bread isnt the same#gaz once got idâd for his energy drink and was so flattered#simon ghost riley#call of duty#cod#call of duty modern warfare#kyle gaz garrick#john soap mactavish#john price#headcanon
221 notes
¡
View notes
Text
âď¸ | automatons
#stim#stimmy#gif#stimboard#stimblr#gifs#gifset#machinery#machine#automatons#automation#robot#robots#mechanics#mechanical#mechanisms#art
96 notes
¡
View notes
Text
"The Writing Boy"

In 1774 AD, during the reign of Louis XVI (1754-1793), Swiss watchmaker Pierre Jacques Dro (1721-1790) unveiled a remarkable engineering feat that would go down in history as the world's first android or programmed automaton.
Known as "The Writing Boy," this creation appeared at first glance to be a simple wooden doll with a porcelain head, barefoot, and holding a goose feather.
But hidden within this seemingly ordinary toy was a technological marvel, a writing mechanism powered by 6,000 intricate moving parts, making it the first automatic calligrapher.
"The Writing Boy" was a groundbreaking achievement, as it was capable of writing complex sentences, such as "My inventor is Jacques Dro."
The automaton was a product of 20 months of meticulous work by Pierre Jacques Dro, and its debut in Paris stunned the court of King Louis XVI.
The android's ability to perform such an intricate task showcased the high level of craftsmanship and innovation of the time.
This astonishing creation marked a significant milestone in the history of robotics and engineering.
Not only was it the worldâs first programmed android, but it also demonstrated the potential of machines to replicate human actions.
"The Writing Boy" paved the way for future advancements in automation, solidifying Pierre Jacques Droâs legacy as a pioneer in the field of robotics.
#Pierre Jacques Dro#Louis XVI#The Writing Boy#first android#android#programmed automaton#automaton#writing mechanism#technological marvel#robotics#engineering#wooden doll#craftsmanship#innovation#machine#automation#1700s#18th century#Swiss watchmaker#automatic calligrapher
146 notes
¡
View notes
Text

#Norbert Wiener#Wiener#cybernetics#technology#computers#book cover#cover design#science#machines#automation
106 notes
¡
View notes
Text
do you think there is any elitism regarding tattoo application in cyberpunk? cause we see v's tattoo being applied automatically by a ripperdoc but i have to assume that tattoo artist is still a job, if even a dying one that's slowly being replaced by machines. like even today we see artists fighting for their work and the integrity of art in general versus ai, and i know in the tattoo community there have been similar things happening re: robotics/automated application within the last few years (look up BlackDot tattoo) and i just want to know what that community is like in 2077. there must still be old school corner tattoo shops right??
#cyberpunk 2077#genuinely dont know if there is any lore about this anywhere#but i was thinking about dagger's tattoos last night and how i just don't see him sitting down and letting a machine#apply his ink lmao. he's got several stick n pokes and for certain his chin tattoo was done by hand#but hes just such an old school type of person with an aversion to technology i don't think he'd want it done by an automated machine#i also have a tattoo artist oc who works out of a shop by hand and i just like. how common is that?????#did machines successfully choke out the artistry????#and do all ripperdocs have tattoo application knowledge as a baseline now because it just merged into that instead?#or is that specifically just the case with that dude who does v's tattoo whose name i don't rmr now cassius or whatever#and this doesn't even begin to touch on tattoos that have some kind of ingrained tech feature too...light up/glow in the dark#what does that tyger claw tattoo thing do its been too long since i played through rip#why is THERE SO MUCH TO THINK ABOUT???? HELLO?????????
38 notes
¡
View notes
Text
The surprising truth about data-driven dictatorships

Hereâs the âdictatorâs dilemmaâ: they want to block their countryâs frustrated elites from mobilizing against them, so they censor public communications; but they also want to know what their people truly believe, so they can head off simmering resentments before they boil over into regime-toppling revolutions.
These two strategies are in tension: the more you censor, the less you know about the true feelings of your citizens and the easier it will be to miss serious problems until they spill over into the streets (think: the fall of the Berlin Wall or Tunisia before the Arab Spring). Dictators try to square this circle with things like private opinion polling or petition systems, but these capture a small slice of the potentially destabiziling moods circulating in the body politic.
Enter AI: back in 2018, Yuval Harari proposed that AI would supercharge dictatorships by mining and summarizing the public moodâââas captured on social mediaâââallowing dictators to tack into serious discontent and diffuse it before it erupted into unequenchable wildfire:
https://www.theatlantic.com/magazine/archive/2018/10/yuval-noah-harari-technology-tyranny/568330/
Harari wrote that âthe desire to concentrate all information and power in one place may become [dictators] decisive advantage in the 21st century.â But other political scientists sharply disagreed. Last year, Henry Farrell, Jeremy Wallace and Abraham Newman published a thoroughgoing rebuttal to Harari in Foreign Affairs:
https://www.foreignaffairs.com/world/spirals-delusion-artificial-intelligence-decision-making
They argued thatâââlike everyone who gets excited about AI, only to have their hopes dashedâââdictators seeking to use AI to understand the public mood would run into serious training data bias problems. After all, people living under dictatorships know that spouting off about their discontent and desire for change is a risky business, so they will self-censor on social media. Thatâs true even if a person isnât afraid of retaliation: if you know that using certain words or phrases in a post will get it autoblocked by a censorbot, whatâs the point of trying to use those words?
The phrase âGarbage In, Garbage Outâ dates back to 1957. Thatâs how long weâve known that a computer that operates on bad data will barf up bad conclusions. But this is a very inconvenient truth for AI weirdos: having given up on manually assembling training data based on careful human judgment with multiple review steps, the AI industry âpivotedâ to mass ingestion of scraped data from the whole internet.
But adding more unreliable data to an unreliable dataset doesnât improve its reliability. GIGO is the iron law of computing, and you canât repeal it by shoveling more garbage into the top of the training funnel:
https://memex.craphound.com/2018/05/29/garbage-in-garbage-out-machine-learning-has-not-repealed-the-iron-law-of-computer-science/
When it comes to âAIâ thatâs used for decision supportâââthat is, when an algorithm tells humans what to do and they do itâââthen you get something worse than Garbage In, Garbage Outâââyou get Garbage In, Garbage Out, Garbage Back In Again. Thatâs when the AI spits out something wrong, and then another AI sucks up that wrong conclusion and uses it to generate more conclusions.
To see this in action, consider the deeply flawed predictive policing systems that cities around the world rely on. These systems suck up crime data from the cops, then predict where crime is going to be, and send cops to those âhotspotsâ to do things like throw Black kids up against a wall and make them turn out their pockets, or pull over drivers and search their cars after pretending to have smelled cannabis.
The problem here is that âcrime the police detectedâ isnât the same as âcrime.â You only find crime where you look for it. For example, there are far more incidents of domestic abuse reported in apartment buildings than in fully detached homes. Thatâs not because apartment dwellers are more likely to be wife-beaters: itâs because domestic abuse is most often reported by a neighbor who hears it through the walls.
So if your cops practice racially biased policing (I know, this is hard to imagine, but stay with me /s), then the crime they detect will already be a function of bias. If you only ever throw Black kids up against a wall and turn out their pockets, then every knife and dime-bag you find in someoneâs pockets will come from some Black kid the cops decided to harass.
Thatâs life without AI. But now letâs throw in predictive policing: feed your âknives found in pocketsâ data to an algorithm and ask it to predict where there are more knives in pockets, and it will send you back to that Black neighborhood and tell you do throw even more Black kids up against a wall and search their pockets. The more you do this, the more knives youâll find, and the more youâll go back and do it again.
This is what Patrick Ball from the Human Rights Data Analysis Group calls âempiricism washingâ: take a biased procedure and feed it to an algorithm, and then you get to go and do more biased procedures, and whenever anyone accuses you of bias, you can insist that youâre just following an empirical conclusion of a neutral algorithm, because âmath canât be racist.â
HRDAG has done excellent work on this, finding a natural experiment that makes the problem of GIGOGBI crystal clear. The National Survey On Drug Use and Health produces the gold standard snapshot of drug use in America. Kristian Lum and William Isaac took Oaklandâs drug arrest data from 2010 and asked Predpol, a leading predictive policing product, to predict where Oaklandâs 2011 drug use would take place.

[Image ID: (a) Number of drug arrests made by Oakland police department, 2010. (1) West Oakland, (2) International Boulevard. (b) Estimated number of drug users, based on 2011 National Survey on Drug Use and Health]
Then, they compared those predictions to the outcomes of the 2011 survey, which shows where actual drug use took place. The two maps couldnât be more different:
https://rss.onlinelibrary.wiley.com/doi/full/10.1111/j.1740-9713.2016.00960.x
Predpol told cops to go and look for drug use in a predominantly Black, working class neighborhood. Meanwhile the NSDUH survey showed the actual drug use took place all over Oakland, with a higher concentration in the Berkeley-neighboring student neighborhood.
Whatâs even more vivid is what happens when you simulate running Predpol on the new arrest data that would be generated by cops following its recommendations. If the cops went to that Black neighborhood and found more drugs there and told Predpol about it, the recommendation gets stronger and more confident.
In other words, GIGOGBI is a system for concentrating bias. Even trace amounts of bias in the original training data get refined and magnified when they are output though a decision support system that directs humans to go an act on that output. Algorithms are to bias what centrifuges are to radioactive ore: a way to turn minute amounts of bias into pluripotent, indestructible toxic waste.
Thereâs a great name for an AI thatâs trained on an AIâs output, courtesy of Jathan Sadowski: âHabsburg AI.â
And that brings me back to the Dictatorâs Dilemma. If your citizens are self-censoring in order to avoid retaliation or algorithmic shadowbanning, then the AI you train on their posts in order to find out what theyâre really thinking will steer you in the opposite direction, so you make bad policies that make people angrier and destabilize things more.
Or at least, that was Farrell(et al)âs theory. And for many years, thatâs where the debate over AI and dictatorship has stalled: theory vs theory. But now, thereâs some empirical data on this, thanks to the âThe Digital Dictatorâs Dilemma,â a new paper from UCSD PhD candidate Eddie Yang:
https://www.eddieyang.net/research/DDD.pdf
Yang figured out a way to test these dueling hypotheses. He got 10 million Chinese social media posts from the start of the pandemic, before companies like Weibo were required to censor certain pandemic-related posts as politically sensitive. Yang treats these posts as a robust snapshot of public opinion: because there was no censorship of pandemic-related chatter, Chinese users were free to post anything they wanted without having to self-censor for fear of retaliation or deletion.
Next, Yang acquired the censorship model used by a real Chinese social media company to decide which posts should be blocked. Using this, he was able to determine which of the posts in the original set would be censored today in China.
That means that Yang knows that the ârealâ sentiment in the Chinese social media snapshot is, and what Chinese authorities would believe it to be if Chinese users were self-censoring all the posts that would be flagged by censorware today.
From here, Yang was able to play with the knobs, and determine how âpreference-falsificationâ (when users lie about their feelings) and self-censorship would give a dictatorship a misleading view of public sentiment. What he finds is that the more repressive a regime isâââthe more people are incentivized to falsify or censor their viewsâââthe worse the system gets at uncovering the true public mood.
Whatâs more, adding additional (bad) data to the system doesnât fix this âmissing dataâ problem. GIGO remains an iron law of computing in this context, too.
But it gets better (or worse, I guess): Yang models a âcrisisâ scenario in which users stop self-censoring and start articulating their true views (because theyâve run out of fucks to give). This is the most dangerous moment for a dictator, and depending on the dictatorship handles it, they either get another decade or rule, or they wake up with guillotines on their lawns.
But âcrisisâ is where AI performs the worst. Trained on the âstatus quoâ data where users are continuously self-censoring and preference-falsifying, AI has no clue how to handle the unvarnished truth. Both its recommendations about what to censor and its summaries of public sentiment are the least accurate when crisis erupts.
But hereâs an interesting wrinkle: Yang scraped a bunch of Chinese usersâ posts from Twitterâââwhich the Chinese government doesnât get to censor (yet) or spy on (yet)âââand fed them to the model. He hypothesized that when Chinese users post to American social media, they donât self-censor or preference-falsify, so this data should help the model improve its accuracy.
He was rightâââthe model got significantly better once it ingested data from Twitter than when it was working solely from Weibo posts. And Yang notes that dictatorships all over the world are widely understood to be scraping western/northern social media.
But even though Twitter data improved the modelâs accuracy, it was still wildly inaccurate, compared to the same model trained on a full set of un-self-censored, un-falsified data. GIGO is not an option, itâs the law (of computing).
Writing about the study on Crooked Timber, Farrell notes that as the world fills up with âgarbage and noiseâ (he invokes Philip K Dickâs delighted coinage âgubbishâ), âapproximately correct knowledge becomes the scarce and valuable resource.â
https://crookedtimber.org/2023/07/25/51610/
This âprobably approximately correct knowledgeâ comes from humans, not LLMs or AI, and so âthe social applications of machine learning in non-authoritarian societies are just as parasitic on these forms of human knowledge production as authoritarian governments.â
The Clarion Science Fiction and Fantasy Writersâ Workshop summer fundraiser is almost over! I am an alum, instructor and volunteer board member for this nonprofit workshop whose alums include Octavia Butler, Kim Stanley Robinson, Bruce Sterling, Nalo Hopkinson, Kameron Hurley, Nnedi Okorafor, Lucius Shepard, and Ted Chiang! Your donations will help us subsidize tuition for students, making Clarionâââand sf/fâââmore accessible for all kinds of writers.
Libro.fm is the indie-bookstore-friendly, DRM-free audiobook alternative to Audible, the Amazon-owned monopolist that locks every book you buy to Amazon forever. When you buy a book on Libro, they share some of the purchase price with a local indie bookstore of your choosing (Libro is the best partner I have in selling my own DRM-free audiobooks!). As of today, Libro is even better, because itâs available in five new territories and currencies: Canada, the UK, the EU, Australia and New Zealand!
[Image ID: An altered image of the Nuremberg rally, with ranked lines of soldiers facing a towering figure in a many-ribboned soldier's coat. He wears a high-peaked cap with a microchip in place of insignia. His head has been replaced with the menacing red eye of HAL9000 from Stanley Kubrick's '2001: A Space Odyssey.' The sky behind him is filled with a 'code waterfall' from 'The Matrix.']
Image: Cryteria (modified) https://commons.wikimedia.org/wiki/File:HAL9000.svg
CC BY 3.0 https://creativecommons.org/licenses/by/3.0/deed.en
âââ
Raimond Spekking (modified) https://commons.wikimedia.org/wiki/File:Acer_Extensa_5220_-_Columbia_MB_06236-1N_-_Intel_Celeron_M_530_-_SLA2G_-_in_Socket_479-5029.jpg
CC BY-SA 4.0 https://creativecommons.org/licenses/by-sa/4.0/deed.en
âââ
Russian Airborne Troops (modified) https://commons.wikimedia.org/wiki/File:Vladislav_Achalov_at_the_Airborne_Troops_Day_in_Moscow_%E2%80%93_August_2,_2008.jpg
âSoldiers of Russiaâ Cultural Center (modified) https://commons.wikimedia.org/wiki/File:Col._Leonid_Khabarov_in_an_everyday_service_uniform.JPG
CC BY-SA 3.0 https://creativecommons.org/licenses/by-sa/3.0/deed.en
#pluralistic#habsburg ai#self censorship#henry farrell#digital dictatorships#machine learning#dictator's dilemma#eddie yang#preference falsification#political science#training bias#scholarship#spirals of delusion#algorithmic bias#ml#Fully automated data driven authoritarianism#authoritarianism#gigo#garbage in garbage out garbage back in#gigogbi#yuval noah harari#gubbish#pkd#philip k dick#phildickian
833 notes
¡
View notes
Text
Me when I blink and suddenly the clock says 6am haha whoops
#My eyes stick together when I blink but fuck it we ball#batim#batdr#batim kin#bendy and the dark revival#bendy and the ink machine#bendy#the ink demon#ink bendy#ink demon#Thank fuck for automated tags I. Am struggking to type LMFAO#Yes this is that one dog image
92 notes
¡
View notes
Text
we for real need to kill automated phone systems and bring back human operators
#i don't want to speak to a machine!!!!!!#that can't even understand me half the time anyways#the absolute worst automated phone lines are the ones where it is almost impossible to talk to an actual person#or sometimes even straight up impossible#they need to die!!!!!!
26 notes
¡
View notes
Text
Sangheili soldier explaining to sangheili civillian that the Covenant has The Machine that can just surgery your body and destroy all your dignity and honor and then you're just supposed to go on like nothing happened, like you weren't just put in The Machine and subjected to The Shame Of It All?
#i just think the pre-programmed automated surgery machines are fascinating and kind of funny cause they're very like#idk videogame#we fucked up the character but we need them to work better for this next part put them in The Box
28 notes
¡
View notes
Photo
What is an Algorithm in 30 Seconds?
An algorithm is simply a series of instructions.
Think of a recipe: boil water, add pasta, wait, drain, eat. These are steps to follow.
In computer terms, an algorithm is a set of instructions for a computer to execute.
In machine learning, these instructions enable computers to learn from data, making machine learning algorithms unique and powerful.
#artificial intelligence#automation#machine learning#business#digital marketing#professional services#marketing#web design#web development#social media#tech#Technology
69 notes
¡
View notes
Text
TW for Ody kins/fictives maybe Penelope kins/fictives read with caution
In EPIC, Penelope was sitting behind the axes.
She was behind the axes.
The task she gave the suitors?
Shoot through 12 axes cleanly.
I of course, understand, why she wouldâve done thisâsheâd rather die than live without her husband.Â
Still kinda fucks me up that she was ready to abandon Tele (I���m not talking in first person for that one because Iâm not positive that was in my canon, and Iâm JUST talking about EPIC). I guess it was Ancient Greece and kids often went without their mothers, but Tele relied on his mother heavily.Â
^^ Still a great song though (The Challenge/The first half of Hold Them Down))
#I always skip the second half of hold them down#I can barely handle the first half lmao#the challenge#hold them down#Epic the musical#E:tM#EtM sounds like ATM#I dunno at ATM stands for#Automated Tax machine?#I'll google it#The Odyssey#Penelope of Ithaca#Odysseus of Ithaca#Telemachus of Ithaca#Antinous of Ithaca unfortunately#@~musically musical~@
15 notes
¡
View notes
Text
I cannot tolerate interacting with USA supporters.
Blue maga pisses me off more than red, because red voters are too stupidly brainwashed to be blamed for much other than shitting the bed. They're pathetic, but violent. Their danger isn't coming from them being too bright or clever.
When liberals shit the bed, it's on purpose & with all the data about bed shitting fully at their fingertips. They're deeply fucking dangerous. They praise the FBI every day, now. They demand cops, they insist on responsible surrender to and complicity with rapacious neoliberalism. They are worse. They're the trustees of the prison and they are proud of it. They are more than capable of cleverly serving Trump up as their excuse for No End of massacre, pillage, theft, pure malice.
Liberals can, should, and will fuck all the way off, forever.
32 notes
¡
View notes