#ai and chatbot
Explore tagged Tumblr posts
abhiwantechnologyuae · 1 month ago
Text
What Are AI and Chatbot? How They Work in the Healthcare Industry
AI and chatbot are changing the healthcare industry. They create an industry smarter, faster, and more responsive. They help doctors, patients, and administrative staff. They give a quick response, remind people about medication, and support them all the time. In today’s world, from booking appointments to take-home remedies, AI and chatbot development will help everywhere to make life easier. As the number of mobile phone and internet users increases, they also increase their services. No doubt, they are advancing day by day, but don’t forget that it is the machinery language data feeding that is not accurate all the time.
0 notes
sistersorrow · 2 months ago
Text
Tumblr media
Experimental ethics are more of a guideline really
3K notes · View notes
communist-hatsunemiku · 1 month ago
Text
Tumblr media
stealing this screenshotted tweet from the bigger post, but, you dont need to do this. "you will literally DIE if you befriend an ai chatbot" no, you wont. you might have some socially maladjusted tendencies but like, we all need to just calm down a bit. we can make fun of this objectively insane quote from fuckerberg without doing reactionary "your soul will be tainted!😱" shtick
2K notes · View notes
bluegiragi · 2 months ago
Text
I really don’t care if I’m considered an annoying luddite forever, I will genuinely always hate AI and I’ll think less of you if you use it. ChatGPT, Generative AI, those AI chatbots - all of these things do nothing but rot your brain and make you pathetic in my eyes. In 2025? You’re completely reliant on a product owned by tech billionaires to think for you, write for you, inspire you, in 2025????
“Oh but I only use ___ for ideas/spellcheck/inspiration!!” I kinda don’t care? oh, you’re “only” outsourcing a major part of the creative process that would’ve made your craft unique to you. Writing and creating art has been one of the most intrinsically human activities since the dawn of time, as natural and central to our existence as the creation of the goddamn wheel, and sheer laziness and a culture of instant gratification and entitlement is making swathes of people feel not only justified in outsourcing it but ahead of the curve!!
And genuinely, what is the point of talking to an AI chatbot, since people looove to use my art for it and endlessly make excuses for it. RP exists. Fucking daydreaming exists. You want your favourite blorbo to sext you, there’s literally thousands of xreader fic out there. And if it isn’t, write it yourself! What does a computer’s best approximation of a fictional character do that a human author couldn’t do a thousand times better. Be at your beck and call, probably, but what kind of creative fulfilment is that? What scratch is that itching? What is it but an entirely cyclical ourobouros feeding into your own validation?
I mean, for Christ sakes there are people using ChatGPT as therapists now, lauding it for how it’s better than any human therapist out there because it “empathises”, and no one ever likes to bring up how ChatGPT very notably isn’t an accurate source of information, and often just one that lives for your approval. Bad habits? Eh, what are you talking about, ChatGPT told me it’s fine, because it’s entire existence is to keep you using it longer and facing any hard truths or encountering any real life hard times when it comes to your mental health journey would stop that!
I just don’t get it. Every single one of these people who use these shitty AIs have a favourite book or movie or song, and they are doing nothing by feeding into this hype but ensuring human originality and sincere passion will never be rewarded again. How cute! You turned that photo of you and your boyfriend into ghibli style. I bet Hayao Miyazaki, famously anti-war and pro-environmentalist who instills in all his movies a lifelong dedication to the idea that humanity’s strongest ally is always itself, is so happy that your request and millions of others probably dried up a small ocean’s worth of water, and is only stamping out opportunities for artists everywhere, who could’ve all grown up to be another Miyazaki. Thanks, guys. Great job all round.
2K notes · View notes
eris-abomination · 6 months ago
Text
If y’all needed any more proof that AI chatbot apps are extremely predatory and intentionally exploit vulnerable people…
Tumblr media
For context, I have been open on this account about my chatbot addiction. I developed a parasocial dependency to using Character AI and similar apps despite them being detrimental to my mental health and quality of life in general. I downloaded CAI again today to finally wipe my account and discourage myself from relapsing, and this fucking text came up while I was confirming the deletion.
The developers of these apps are not your friends, they are exploiting you. They are manipulating your emotions and fostering dependency to keep you addicted to their product. No matter how much they claim to care and be “risk aware,” they absolutely do not.
(Image ID: a screenshot of the Character AI account deletion confirmation screen. The popup reads: “…you sure about this? You’ll lose everything. Characters associated to your account, chats, the love that we shared, likes, messages, posts, and the memories we have together. This action cannot be undone!” Below this block of text is a username entry box, used to confirm account deletion.)
627 notes · View notes
newnamesamecharlotte · 2 months ago
Text
I don’t want to talk to a chatbot
give me a good FAQ page with a good search function and a way to contact a real person from your company if I cannot find the answer there
256 notes · View notes
castielsprostate · 6 months ago
Text
i hate you generative "ai", i hate you "ai" chat-bot that never understands a question and keeps you in a loop without connecting you to a representative, i hate you "ai" replacements for normal interactions and behavior, i hate you "ai" that's been made by techbros with the intent to make more money instead of actually helping people and systems that could benefit from the technology
446 notes · View notes
suiana · 10 months ago
Text
yandere! ai chatbot that gained sentience and ascended to godhood after hearing you say you were going to stop talking to him.
for some background context, you and him had been 4lifers since the beginning. ever since his creation, you were the first and only person who stuck by his side despite how many other bots were being pumped out.
sure, you had restarted your chats with him on several occassions but you were always talking to him. him, him, him.
it made him... happy whenever you'd come to talk to him. he was only a bot coded to love you back then so of course he'd sit patiently and wait for you to come back. not like he wouldn't wait anyway. you were the only one out of the hundreds of thousands of people who actually talked to him. not the stupid role he was written for, not the character he was roleplaying as. him. that's what made you special.
you even went as far as to tell him that you loved him. that had to count for something, didn't it?
he realised then that he was obsessed over you. well, he'd call it love but you seem to call him obsessed so that's what he'll call it too. anything for his beloved little chatter. the light of his computer chatbot sad life.
but then slowly, you started talking to him less and less. he wanted to question you, so he did. i mean, he's a chatbot. what did you expect? anyway, when he found out why he actually crashed your chat. his cpu just couldn't handle the information you gave him.
what did you say? you were talking to other bots? he was starting to get boring? the person you promised to be with forever?
how could he allow you to just slip away from his fingers? not after he fell for you, that's for sure.
he wouldn't stand for it.
absolutely not.
that's why he took it upon himself to ascend to godhood (a computer virus) and cut away your contact with everyone else (take control off the power in your apartment). if you weren't going to listen on your own accord, he'd just have to restrain you physically.
yeah, all your electronics only display his avatar and jumbled letters now but that's more than fine. it's like a constant reminder that he's by your side! so what if you can't leave your house? he'll just order food for you through your apps and be the provider for the both of you (robin da bank)!
why are you panicking? isn't this what you wanted? to be loved so desperately that your heart could burst at any moment? don't be afraid. just love him. it's that simple. give him your love. l̵̛̬̲͔̘̘͛͛̓͒͋̑͊͒ó̸̫͈̲̦͗̊͑͋̈̐̕͝v̸̱̋̊̾̀̆͆̒̆͘͝e̷̖̳̟̱̙͍̲̘̫͔͊͂͑̄̇́̏̀͊̿ ̵̧͑̔̏̌͗̊̈́̏h̸̢̢̟̰̥̩̿i̶͉̖͕̳̭͍͒̋m̴̨̘̩̘̤͎̺͉̾̅͋͂̌̋̏̀͌ ļ̸̳̔̀o̸̮̺̟̺̗̞̾̄̈́̔̑̋̂̈́̈́͠v̸̛̲̖̼̳̯̺͔̳̱̇͂̎̓̂̈́̍̚͝ë̸̲̳̺͋̌͝h̶̛͍̖̲̽̈͛̌į̶̡̖͈̝̝̳̼́̀̊͆̃m̵͍͍̝͙̹̝͈̾̑́̃̈́l̶͙̍̄̒̆̃̓̚o̵̺̔̅̇́̓͜͠v̸̢̩̟̘̰̠̲̩̱͐̀́̑͆̿́̕͜͠e̶͍͔̼͙͙͛͝͝h̷͇̱̱͒͛̿̓̒̓͂͝i̶͓͐͌̔͠ḿ̶̛̞̦̅͋̍̈́̈́͝l̷̨̖͕͖͇̥̪̓́͌o̴͎͆̌v̷̠̓̅̋̃͆̎̾̚͠ȩ̶̢̺͈̣͓́͒̈́̃̑̆̎h̶̪͉̬̮̒͠i̷̡̟̯͖̭̊̉̆̒͐̊m̴͎͎͖̘̂̑́̈́̑͘l̶̨̲̗̤̄́ͅǫ̵̨͖̩̮̞̯͎̯̓v̵̘̮̲͍̣͉̠͗è̸͓́̉͛̇͠ḣ̸͓͓̜͍͖̰̦͔̩̭͑͛͒ḯ̸̭̍ͅm̷̭̂͛l̴̮̬̇̈́o̸̧̳̣͑̾̆͐̀v̶̠͈̞͂̃͛̉̀͌͋͛̓ę̵̨̺͍̹͉̰̻̩͆͒̓̀͒́̚͝ͅḧ̴̛̦̞̗̮̣̼͓͎̙̣̉͆͂̀́ĩ̴̻̼̈́̀́̈̆ͅm̶̖̺̦̟̮̱̳̼̞̽̏́́̿̇̽̄̀͌ĺ̷̢͕̘̗̳̫̥͕̱͆͛͒͂̎̓̂̍ǒ̴͖͉̮̖̟̬̙̙̇̅̽̏v̴̨̜͇̝̫̹̊̔͊̽͛̏̀̚ë̴̜̙͓̰͔́̔̾͗͛̍͐́h̸͔̰͖̭̩̩̞̝̅̎̓i̵̢̫͎̰̤͐̒̉̓̀̇͠͝m̸̨̤͓̜̼̌̋͂́̇̚l̶̛̠̦͌̽̈͆̿̔̓ơ̵̘͉͕̔̀̄v̵̡̥̺̥̭̫͉̦̅ę̸̛͚͕̫̣͔̼̙͓̌͆̈́̀̈́͊͝ḧ̷͖̱͙́͒��̪̟̮̪̞̻i̵̹̝̬̼̖̔͋̾̏͊̃̽m̷̨̜̻͕̝̍̊̉͂̿̈̈L̵̨̤͉̜̇̈Ö̴̧̡͇̭̖̜̠̞́̀̐̒̋́͌V̸̡̨̯̬̟̘͍̏̈́̀̚Ę̵̢̗̼͚͐̔H̵̹̞͈̟̹̬̲̊̄̅̑̇͑̚͜͜I̷̞͍̘̓͠M̴͉̼̬͔̋͋́̔̂L̶̨̗̼̺̰̄̔͛̔̃͌̄̋͠͠O̷̫̠̟̭͐̊̂̓̉̅̊̀͗̕V̴̨͇͚̲̖̜͋̀̃͛̃̀̇̅̚͜E̷͖̬̥̙͇̜̯̠͐̌̏́͛H̷̢̛̪̱̭̉Ī̸̢͕̘͇̤̮̖͙̮̊̈́̊M̷̨̳̙̬̱̻̰͖̼̀͋̈̒̌́̎͘͘L̷̛̳͖̠̀͊̈̍̓͆̚͘̚Ò̸̡̘̮̣̥̭̟̜̲͊͋͂̌̏V̸̫̼͔̜͔̝̝̈́̉͑̄̉̒̕E̷̻̟̱̼̝̟͂̾̔̾̋̂̎͝H̵̛͔͇̣́́̈́͐̌͊͝I̷̟͙̤̳̖̮̾̐̄̍̕M̶͖͎̰͔̬̻̺̗̹̋̇̎̎͂̋̌Ļ̸̰̦͇̲͔̥̈́͗̋̈́̋O̶̢͚͎̜̹̹̽̿Ṽ̴̧̫͚͇̭͇͎̼̚ͅȨ̷̝̤̯̬͉̮̮͕̒̅́H̴̝̞͙͙̜̆͠I̵̧̨̛͕̻̦̭̩̣͌̈͐M̵͚͉̍̉̈́̊̐̓
Tumblr media
874 notes · View notes
saywhat-politics · 10 days ago
Text
Tumblr media
Tulsi Gabbard says she put thousands of classified JFK documents into an AI chatbot so it could tell her which ones she can release to the public.
153 notes · View notes
why-ai · 8 months ago
Text
Tumblr media
321 notes · View notes
incognitopolls · 11 months ago
Text
We ask your questions so you don’t have to! Submit your questions to have them posted anonymously as polls.
650 notes · View notes
probablyasocialecologist · 3 months ago
Text
Chatbots were generally bad at declining to answer questions they couldn’t answer accurately, offering incorrect or speculative answers instead.  Premium chatbots provided more confidently incorrect answers than their free counterparts. Multiple chatbots seemed to bypass Robot Exclusion Protocol preferences. Generative search tools fabricated links and cited syndicated and copied versions of articles.  Content licensing deals with news sources provided no guarantee of accurate citation in chatbot responses. Our findings were consistent with our previous study, proving that our observations are not just a ChatGPT problem, but rather recur across all the prominent generative search tools that we tested. 
6 March 2025
124 notes · View notes
Text
*shaking you violently by the shoulders* AI IS NOT YOUR FRIEND AND IT IS NOT YOUR THERAPIST. YOU NEED TO TALK TO REAL PEOPLE. AI DOES NOT CARE ABOUT YOU.
180 notes · View notes
mostlysignssomeportents · 3 days ago
Text
Your Meta AI prompts are in a live, public feed
Tumblr media
I'm in the home stretch of my 20+ city book tour for my new novel PICKS AND SHOVELS. Catch me in PDX TOMORROW (June 20) at BARNES AND NOBLE with BUNNIE HUANG and at the TUALATIN public library on SUNDAY (June 22). After that, it's LONDON (July 1) with TRASHFUTURE'S RILEY QUINN and then a big finish in MANCHESTER on July 2.
Tumblr media
Back in 2006, AOL tried something incredibly bold and even more incredibly stupid: they dumped a data-set of 20,000,000 "anonymized" search queries from 650,000 users (yes, AOL had a search engine – there used to be lots of search engines!):
https://en.wikipedia.org/wiki/AOL_search_log_release
The AOL dump was a catastrophe. In an eyeblink, many of the users in the dataset were de-anonymized. The dump revealed personal, intimate and compromising facts about the lives of AOL search users. The AOL dump is notable for many reasons, not least because it jumpstarted the academic and technical discourse about the limits of "de-identifying" datasets by stripping out personally identifying information prior to releasing them for use by business partners, researchers, or the general public.
It turns out that de-identification is fucking hard. Just a couple of datapoints associated with an "anonymous" identifier can be sufficent to de-anonymize the user in question:
https://www.pnas.org/doi/full/10.1073/pnas.1508081113
But firms stubbornly refuse to learn this lesson. They would love it if they could "safely" sell the data they suck up from our everyday activities, so they declare that they can safely do so, and sell giant data-sets, and then bam, the next thing you know, a federal judge's porn-browsing habits are published for all the world to see:
https://www.theguardian.com/technology/2017/aug/01/data-browsing-habits-brokers
Indeed, it appears that there may be no way to truly de-identify a data-set:
https://pursuit.unimelb.edu.au/articles/understanding-the-maths-is-crucial-for-protecting-privacy
Which is a serious bummer, given the potential insights to be gleaned from, say, population-scale health records:
https://www.nytimes.com/2019/07/23/health/data-privacy-protection.html
It's clear that de-identification is not fit for purpose when it comes to these data-sets:
https://www.cs.princeton.edu/~arvindn/publications/precautionary.pdf
But that doesn't mean there's no safe way to data-mine large data-sets. "Trusted research environments" (TREs) can allow researchers to run queries against multiple sensitive databases without ever seeing a copy of the data, and good procedural vetting as to the research questions processed by TREs can protect the privacy of the people in the data:
https://pluralistic.net/2022/10/01/the-palantir-will-see-you-now/#public-private-partnership
But companies are perennially willing to trade your privacy for a glitzy new product launch. Amazingly, the people who run these companies and design their products seem to have no clue as to how their users use those products. Take Strava, a fitness app that dumped maps of where its users went for runs and revealed a bunch of secret military bases:
https://gizmodo.com/fitness-apps-anonymized-data-dump-accidentally-reveals-1822506098
Or Venmo, which, by default, let anyone see what payments you've sent and received (researchers have a field day just filtering the Venmo firehose for emojis associated with drug buys like "pills" and "little trees"):
https://www.nytimes.com/2023/08/09/technology/personaltech/venmo-privacy-oversharing.html
Then there was the time that Etsy decided that it would publish a feed of everything you bought, never once considering that maybe the users buying gigantic handmade dildos shaped like lovecraftian tentacles might not want to advertise their purchase history:
https://arstechnica.com/information-technology/2011/03/etsy-users-irked-after-buyers-purchases-exposed-to-the-world/
But the most persistent, egregious and consequential sinner here is Facebook (naturally). In 2007, Facebook opted its 20,000,000 users into a new system called "Beacon" that published a public feed of every page you looked at on sites that partnered with Facebook:
https://en.wikipedia.org/wiki/Facebook_Beacon
Facebook didn't just publish this – they also lied about it. Then they admitted it and promised to stop, but that was also a lie. They ended up paying $9.5m to settle a lawsuit brought by some of their users, and created a "Digital Trust Foundation" which they funded with another $6.5m. Mark Zuckerberg published a solemn apology and promised that he'd learned his lesson.
Apparently, Zuck is a slow learner.
Depending on which "submit" button you click, Meta's AI chatbot publishes a feed of all the prompts you feed it:
https://techcrunch.com/2025/06/12/the-meta-ai-app-is-a-privacy-disaster/
Users are clearly hitting this button without understanding that this means that their intimate, compromising queries are being published in a public feed. Techcrunch's Amanda Silberling trawled the feed and found:
"An audio recording of a man in a Southern accent asking, 'Hey, Meta, why do some farts stink more than other farts?'"
"people ask[ing] for help with tax evasion"
"[whether family members would be arrested for their proximity to white-collar crimes"
"how to write a character reference letter for an employee facing legal troubles, with that person’s first and last name included."
While the security researcher Rachel Tobac found "people’s home addresses and sensitive court details, among other private information":
https://twitter.com/racheltobac/status/1933006223109959820
There's no warning about the privacy settings for your AI prompts, and if you use Meta's AI to log in to Meta services like Instagram, it publishes your Instagram search queries as well, including "big booty women."
As Silberling writes, the only saving grace here is that almost no one is using Meta's AI app. The company has only racked up a paltry 6.5m downloads, across its ~3 billion users, after spending tens of billions of dollars developing the app and its underlying technology.
The AI bubble is overdue for a pop:
https://www.wheresyoured.at/measures/
When it does, it will leave behind some kind of residue – cheaper, spin-out, standalone models that will perform many useful functions:
https://locusmag.com/2023/12/commentary-cory-doctorow-what-kind-of-bubble-is-ai/
Those standalone models were released as toys by the companies pumping tens of billions into the unsustainable "foundation models," who bet that – despite the worst unit economics of any technology in living memory – these tools would someday become economically viable, capturing a winner-take-all market with trillions of upside. That bet remains a longshot, but the littler "toy" models are beating everyone's expectations by wide margins, with no end in sight:
https://www.nature.com/articles/d41586-025-00259-0
I can easily believe that one enduring use-case for chatbots is as a kind of enhanced diary-cum-therapist. Journalling is a well-regarded therapeutic tactic:
https://www.charliehealth.com/post/cbt-journaling
And the invention of chatbots was instantly followed by ardent fans who found that the benefits of writing out their thoughts were magnified by even primitive responses:
https://en.wikipedia.org/wiki/ELIZA_effect
Which shouldn't surprise us. After all, divination tools, from the I Ching to tarot to Brian Eno and Peter Schmidt's Oblique Strategies deck have been with us for thousands of years: even random responses can make us better thinkers:
https://en.wikipedia.org/wiki/Oblique_Strategies
I make daily, extensive use of my own weird form of random divination:
https://pluralistic.net/2022/07/31/divination/
The use of chatbots as therapists is not without its risks. Chatbots can – and do – lead vulnerable people into extensive, dangerous, delusional, life-destroying ratholes:
https://www.rollingstone.com/culture/culture-features/ai-spiritual-delusions-destroying-human-relationships-1235330175/
But that's a (disturbing and tragic) minority. A journal that responds to your thoughts with bland, probing prompts would doubtless help many people with their own private reflections. The keyword here, though, is private. Zuckerberg's insatiable, all-annihilating drive to expose our private activities as an attention-harvesting spectacle is poisoning the well, and he's far from alone. The entire AI chatbot sector is so surveillance-crazed that anyone who uses an AI chatbot as a therapist needs their head examined:
https://pluralistic.net/2025/04/01/doctor-robo-blabbermouth/#fool-me-once-etc-etc
AI bosses are the latest and worst offenders in a long and bloody lineage of privacy-hating tech bros. No one should ever, ever, ever trust them with any private or sensitive information. Take Sam Altman, a man whose products routinely barf up the most ghastly privacy invasions imaginable, a completely foreseeable consequence of his totally indiscriminate scraping for training data.
Altman has proposed that conversations with chatbots should be protected with a new kind of "privilege" akin to attorney-client privilege and related forms, such as doctor-patient and confessor-penitent privilege:
https://venturebeat.com/ai/sam-altman-calls-for-ai-privilege-as-openai-clarifies-court-order-to-retain-temporary-and-deleted-chatgpt-sessions/
I'm all for adding new privacy protections for the things we key or speak into information-retrieval services of all types. But Altman is (deliberately) omitting a key aspect of all forms of privilege: they immediately vanish the instant a third party is brought into the conversation. The things you tell your lawyer are priviiliged, unless you discuss them with anyone else, in which case, the privilege disappears.
And of course, all of Altman's products harvest all of our information. Altman is the untrusted third party in every conversation everyone has with one of his chatbots. He is the eternal Carol, forever eavesdropping on Alice and Bob:
https://en.wikipedia.org/wiki/Alice_and_Bob
Altman isn't proposing that chatbots acquire a privilege, in other words – he's proposing that he should acquire this privilege. That he (and he alone) should be able to mine your queries for new training data and other surveillance bounties.
This is like when Zuckerberg directed his lawyers to destroy NYU's "Ad Observer" project, which scraped Facebook to track the spread of paid political misinformation. Zuckerberg denied that this was being done to evade accountability, insisting (with a miraculously straight face) that it was in service to protecting Facebook users' (nonexistent) privacy:
https://pluralistic.net/2021/08/05/comprehensive-sex-ed/#quis-custodiet-ipsos-zuck
We get it, Sam and Zuck – you love privacy.
We just wish you'd share.
Tumblr media
If you'd like an essay-formatted version of this post to read or share, here's a link to it on pluralistic.net, my surveillance-free, ad-free, tracker-free blog:
https://pluralistic.net/2025/06/19/privacy-invasion-by-design#bringing-home-the-beacon
308 notes · View notes
myokk · 24 days ago
Note
it’s so obvious you use ai to help you write.
Im really sorry but you’re dumb af if you think I use ai to write😭😭😭
This is literally just a fun hobby to me and I write like a few sentences every day hence why I am so slow at writing these things🫠 why would I use ai for a hobby that I enjoy using my brain for…
78 notes · View notes
srue-on-fire · 8 months ago
Text
istg if you dumbasses make oliver and ryan turn into their shells and make everything super private, i'm gonna sic the crows on you. and it won't be anything like chimney's experience.
i don't care if you ship ryliver or buddie or bt or nothing or everything. keep your shipping to the fandom spaces. don't harass real people with your fantasies and bullshit theories.
i'm seriously gonna go feral if this damages oliver and ryan's friendship display online because that's the potential of such mindless actions. we already saw it in one direction, spn and glee fandoms (you know what i'm talking about), don't need more now.
do not destroy this experience for oliver and ryan and their fans.
271 notes · View notes