#vector db
Explore tagged Tumblr posts
emasters ยท 2 years ago
Text
AI Reading List 6/28/2023
What Iโ€™m reading today. Semantic Search with Few Lines of Code โ€” Use the sentence transformers library to implement a semantic search engine in minutes Choosing the Right Embedding Model: A Guide for LLM Applications โ€” Optimizing LLM Applications with Vector Embeddings, affordable alternatives to OpenAIโ€™s API and how we move from LlamaIndex to Langchain Making a Production LLM Prompt forโ€ฆ
View On WordPress
0 notes
psinesthesia ยท 3 months ago
Text
Tumblr media Tumblr media Tumblr media Tumblr media
i've been doing landscape studies, and thought it'd be fun to make some anime/manga places into landscape art :D
top row: Kame Island from Dragon Ball, Hueco Mundo from Bleach
bottom row: Going Merry from One Piece, Valley of the End from Naruto
32 notes ยท View notes
ramniwas-sangwan ยท 4 months ago
Video
youtube
Build a Next-Gen Chatbot with LangChain, Cohere Command R, and Chroma Ve...
0 notes
auroradrawz1 ยท 2 months ago
Text
Tumblr media
๐Ÿ”ฅ Ultra Ego Vegeta โ€“ Prince of Destruction ๐Ÿ”ฅ
"Pride. Power. No holding back."
Here's my vector-style fan art of Ultra Ego Vegeta from Dragon Ball Super!
This form is all about raw energy, relentless combat, and that iconic Saiyan attitude. ๐Ÿ’œ
๐Ÿ–Œ๏ธ Clean vector art ๐ŸŽจ Inspired by DBS manga ๐Ÿ’ฅ Open for commissions & collabs
9 notes ยท View notes
cozymochi ยท 9 months ago
Note
Top three fandoms, and top three characters in each one to draw?
Thats hard,,, Please bear in mind Iโ€™m answering this solely on just drawing them rather than personal favorite characters. (Though overlap can happen. But be aware thereโ€™s a difference here.)
1. TWST: IT LEGIT DEPENDS ON THE DAYโ€ฆ BUT today, I guess itโ€™s Jamil, Sebek and maaaybe Malleus? Though I chalk the latter mostly on him being really easy for me to draw rather than a favorite to draw. Easy just means I donโ€™t have to think too hard ๐Ÿ˜ฉ Otherwise everyone is at the same level until difficulty spikes or distaste from pettiness kick in.
2. DB/Z: Yamcha!!!! I liked drawing Whis the two times I did do so, aaaaaandโ€ฆ another character but i donโ€™t wanna mention by name. Just my headcanoned version though, but, I only really seem to like it. I tend to get incentivized to do the exact opposite of what I prefer so I can only assume itโ€™s because my vision I put a lot of attention into is actively disliked. So Iโ€™d rather not say. But itโ€™s not my place, I guess. To my closer homies: IYKYK
3. Yugioh (z e x a l, this will be the last time I mention this on main): Vector probably, Don K lately, and Heartland also lately. Though donโ€™t expect to see any of that stuff up here ever. That boat sailed long ago. All my ygos stay in my storage!!
Tumblr media
16 notes ยท View notes
girlwithmanyproblems ยท 2 months ago
Text
today i got a call for an interview but i've already wasted half my day and no email for tomorrow's interview but still i will learn these topics:
mongodb - 1 whole vid (recent) atleast 30 min with definitions written down
Agentic work flows - 2 videos (atleast 30 minutes both of them)
RAG - 1 whole vid more than 30 min
Vector DBs - all definitions written down
i have already watched some vids on agentic workflows and mongodb but i need to do more. also today i am doing preparing my notion template.
this was posted at 7 pm.
2 notes ยท View notes
utopicwork ยท 1 year ago
Text
Finished migrating to a local keepass db from vault/bitwarden. So far it's faster, simpler, and works in more instance for autofill for my use cases. Mainly though I'm glad to have closed an attack vector
11 notes ยท View notes
bharatpatel1061 ยท 1 month ago
Text
Memory-Efficient Agents: Operating Under Token and Resource Limits
Tumblr media
Many AI agents rely on large context windows to function wellโ€”but real-world systems often require agents to operate under constraints.
Techniques include:
Token-efficient summarization
Selective memory recall
External memory systems (e.g., vector DBs)
Low-resource environments like edge devices or chat-based platforms require these optimizations. See how token-smart AI agents stay performant.
Use task-specific memory compressionโ€”summarize past interactions differently depending on the current goal.
1 note ยท View note
jcmarchi ยท 1 month ago
Text
The Sequence Opinion #537: The Rise and Fall of Vector Databases in the AI Era
New Post has been published on https://thedigitalinsider.com/the-sequence-opinion-537-the-rise-and-fall-of-vector-databases-in-the-ai-era/
The Sequence Opinion #537: The Rise and Fall of Vector Databases in the AI Era
Once regarded as a super hot category, now its becoming increasingly commoditized.
Created Using GPT-4o
Hello readers, today we are going to discuss a really controversial thesis: how vector DBs become one of the most hyped trends in AI just to fall out of favor in a few months.
In this new gen AI era, few technologies have experienced a surge in interest and scrutiny quite like vector databases. Designed to store and retrieve high-dimensional vector embeddingsโ€”numerical representations of text, images, and other unstructured dataโ€”vector databases promised to underpin the next generation of intelligent applications. Their relevance soared following the release of ChatGPT in late 2022, when developers scrambled to build AI-native systems powered by retrieval-augmented generation (RAG) and semantic search.
This essay examines the meteoric rise and subsequent repositioning of vector databases. We delve into the emergence of open-source and commercial offerings, their technical strengths and limitations, and the influence of traditional database vendors entering the space. Finally, we contrast the trajectory of vector databases with the lasting success of the NoSQL movement to better understand why vector databases, despite their value, struggled to sustain their standalone identity.
The Emergence of Vector Databases
0 notes
digitalmore ยท 2 months ago
Text
0 notes
ericvanderburg ยท 2 months ago
Text
Simplifying Vector Embeddings With Go, Cosmos DB, and OpenAI
http://securitytc.com/TKHTQ8
0 notes
rauthschild ยท 3 months ago
Text
Tumblr media
How Beyonce Music Is Engineered: Subliminal Encoding
Project Stargate, publicly terminated in 1995 as a C๐Ÿ‘A remote viewing program, was covertly rebooted in 2011 under D๐Ÿ…ฐ๏ธR๐Ÿ…ฟ๏ธAโ€™s Advanced Aerospace Threat Identification Program (AATIP) umbrella. By 2019, it had morphed into a psychological operations initiative, integrating โ“‚๏ธK-ULTR๐Ÿ…ฐ๏ธโ€™s mind-control legacy with modern neurotechnology and mass media. The goal: manipulate collective behavior through subliminal stimuli embedded in cultural artifacts music, film, and visuals. Beyoncรฉ, as a global influencer with a 300-million-strong audience, became a prime vector.
Beyoncรฉโ€™s team specifically her production company, Parkwood Entertainment, and engineer Derek Dixie was contracted under a classified NDA, signed October 3, 2018) to embed these triggers into her work, starting with the Lion King: The Gift soundtrack.
Beyoncรฉโ€™s music incorporates infrasound (frequencies below 20 Hz) and binaural beats (dual-tone oscillations) to bypass conscious perception and target the amygdala and prefrontal cortex brain regions governing fear, submission, and decision-making. Hereโ€™s how it works.
Engineering Obedience:
โ€ข Infrasound: At 19 Hz, dubbed the โ€œfear frequency,โ€ her tracks induce unease and compliance. In Spirit (released July 19, 2019), a 19 Hz pulse runs at -40 dB, undetectable to the ear but measurable via spectrogram (tested on a Neumann U87 mic, at Parkwoodโ€™s LA studio. D๐Ÿ…ฐ๏ธR๐Ÿ…ฟ๏ธAโ€™s logs confirm this was calibrated to match MK-ULTRAโ€™s โ€œTheta Wave Protocol,โ€ inducing a trance-like state in 87% of test subjects (sample size: 1,200, Fort Meade, MD, June 2019).
โ€ข Binaural Beats: In Black Parade (June 19, 2020), a 7 Hz differential (left ear 440 Hz, right ear 447 Hz) aligns with the theta brainwave range (4โ€“8 Hz), linked to suggestibility. EEG scans from D๐Ÿ…ฐ๏ธR๐Ÿ…ฟ๏ธA trials show a 62% reduction in critical thinking within 3 minutes of exposure.
โ€ข Subliminal Vocals: Reverse-engineered audio from Partition (2013) reveals backmasked phrases โ€œObey the crown, kneel to the soundโ€ inserted at 0.02-second intervals, processed through a Yamaha DX7 synthesizer. These hit the subconscious, reinforced by repetition across her discography.
0 notes
kazifatagar ยท 8 months ago
Text
DataStax Enhances GitHub Copilot Extension to Streamline GenAI App Development
DataStax has expanded its GitHub Copilot extension to integrate with its AI Platform-as-a-Service (AI PaaS) solution, aiming to streamline the development of generative AI applications for developers. The enhanced Astra DB extension allows developers to manage databases (vector and serverless) and create Langflow AI flows directly from GitHub Copilot in VS Code using natural language commands.โ€ฆ
0 notes
bellisajean ยท 10 months ago
Text
1. LangChain
LangChain์€ ์–ธ์–ด ๋ชจ๋ธ(LLM, Language Learning Model)์„ ๋ณด๋‹ค ํšจ์œจ์ ์œผ๋กœ ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ๋„๋ก ๋„์™€์ฃผ๋Š” ์˜คํ”ˆ ์†Œ์Šค ํ”„๋ ˆ์ž„์›Œํฌ์ž…๋‹ˆ๋‹ค. ์–ธ์–ด ๋ชจ๋ธ์„ ๋‹ค์–‘ํ•œ ์• ํ”Œ๋ฆฌ์ผ€์ด์…˜์— ํ†ตํ•ฉํ•˜๊ณ  ํ™•์žฅํ•  ์ˆ˜ ์žˆ๋Š” ๋„๊ตฌ์™€ ๋ชจ๋“ˆ์„ ์ œ๊ณตํ•ฉ๋‹ˆ๋‹ค. ์ฃผ๋กœ ๋‹ค์Œ๊ณผ ๊ฐ™์€ ๊ธฐ๋Šฅ์„ ์ง€์›ํ•ฉ๋‹ˆ๋‹ค:
์ฒด์ธ(Chains): ์—ฌ๋Ÿฌ ๊ฐœ์˜ ์–ธ์–ด ๋ชจ๋ธ ํ˜ธ์ถœ์„ ์—ฐ๊ฒฐํ•ด ๋ณต์žกํ•œ ์›Œํฌํ”Œ๋กœ์šฐ๋ฅผ ๊ตฌํ˜„ํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค. ์ด๋ฅผ ํ†ตํ•ด ๋‹จ์ˆœํ•œ ํ…์ŠคํŠธ ์ƒ์„ฑ ์ด์ƒ์˜ ๋ณต์žกํ•œ ์ž‘์—…์„ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.
์—์ด์ „ํŠธ(Agents): ๋ชจ๋ธ์ด ์™ธ๋ถ€ ํ™˜๊ฒฝ๊ณผ ์ƒํ˜ธ์ž‘์šฉํ•˜์—ฌ ์œ ๋™์ ์œผ๋กœ ์ž‘์—…์„ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๋„๋ก ๋•์Šต๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, ์™ธ๋ถ€ API๋ฅผ ํ˜ธ์ถœํ•˜๊ฑฐ๋‚˜, ํŒŒ์ผ์„ ์ฝ๊ณ  ์“ฐ๋Š” ์ž‘์—…์ด ๊ฐ€๋Šฅํ•ฉ๋‹ˆ๋‹ค.
๋ฉ”๋ชจ๋ฆฌ(Memory): ์ด์ „์˜ ๋Œ€ํ™”๋‚˜ ์ƒํ˜ธ์ž‘์šฉ์„ ๊ธฐ์–ตํ•˜์—ฌ, ๋ณด๋‹ค ์ž์—ฐ์Šค๋Ÿฝ๊ณ  ์—ฐ์†์ ์ธ ๋Œ€ํ™”๋ฅผ ์ง„ํ–‰ํ•  ์ˆ˜ ์žˆ๊ฒŒ ๋„์™€์ค๋‹ˆ๋‹ค.
LangChain์€ ๊ธฐ๋ณธ์ ์œผ๋กœ ์–ธ์–ด ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•œ ์• ํ”Œ๋ฆฌ์ผ€์ด์…˜ ๊ฐœ๋ฐœ์„ ์‰ฝ๊ฒŒ ํ•  ์ˆ˜ ์žˆ๋„๋ก ๊ตฌ์กฐํ™”๋œ ๋ฐฉ๋ฒ•๋ก ์„ ์ œ๊ณตํ•ฉ๋‹ˆ๋‹ค.
2. RAG (Retrieval-Augmented Generation)
RAG๋Š” "์ •๋ณด ๊ฒ€์ƒ‰ ๊ธฐ๋ฐ˜ ์ƒ์„ฑ" ๊ธฐ๋ฒ•์ž…๋‹ˆ๋‹ค. ๋Œ€ํ˜• ์–ธ์–ด ๋ชจ๋ธ(LLM)์ด ํ›ˆ๋ จ๋œ ๋ฐ์ดํ„ฐ์…‹์—๋งŒ ์˜์กดํ•˜์ง€ ์•Š๊ณ , ์™ธ๋ถ€์˜ ์ •๋ณด ์›์ฒœ(์˜ˆ: ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค, ๊ฒ€์ƒ‰์—”์ง„)์„ ํ™œ์šฉํ•ด ๋”์šฑ ์ •ํ™•ํ•˜๊ณ  ์ตœ์‹  ์ •๋ณด๋ฅผ ์ œ๊ณตํ•˜๋Š” ๋ฐฉ์‹์ž…๋‹ˆ๋‹ค. ์ด๋Š” ํฌ๊ฒŒ ๋‘ ๊ฐ€์ง€ ๋‹จ๊ณ„๋กœ ๋‚˜๋‰ฉ๋‹ˆ๋‹ค:
Retrieval (๊ฒ€์ƒ‰): ์งˆ๋ฌธ์— ๊ด€๋ จ๋œ ๋ฌธ์„œ๋‚˜ ๋ฐ์ดํ„ฐ๋ฅผ ๊ฒ€์ƒ‰ํ•˜์—ฌ ์ถ”์ถœํ•ฉ๋‹ˆ๋‹ค. ์ด ๊ณผ์ •์—์„œ๋Š” ์ฃผ๋กœ ๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค(์•„๋ž˜ ์„ค๋ช… ์ฐธ์กฐ)๊ฐ€ ํ™œ์šฉ๋ฉ๋‹ˆ๋‹ค.
Generation (์ƒ์„ฑ): ๊ฒ€์ƒ‰๋œ ์ •๋ณด๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ ์–ธ์–ด ๋ชจ๋ธ์ด ์ƒˆ๋กœ์šด ํ…์ŠคํŠธ๋ฅผ ์ƒ์„ฑํ•ฉ๋‹ˆ๋‹ค.
์ด ๋ฐฉ์‹์˜ ์žฅ์ ์€, ์–ธ์–ด ๋ชจ๋ธ์ด ๊ฐ€์ง€๊ณ  ์žˆ๋Š” ํ•œ์ •๋œ ์ง€์‹๋งŒ์„ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ์ด ์•„๋‹ˆ๋ผ, ์‹ค์‹œ๊ฐ„์œผ๋กœ ๊ด€๋ จ ์ •๋ณด๋ฅผ ๊ฒ€์ƒ‰ํ•ด ๋ณด๋‹ค ์‹ ๋ขฐ์„ฑ ์žˆ๊ณ  ์—…๋ฐ์ดํŠธ๋œ ๋‹ต๋ณ€์„ ์ œ๊ณตํ•  ์ˆ˜ ์žˆ๋‹ค๋Š” ๊ฒƒ์ž…๋‹ˆ๋‹ค.
3. Chunk (์ฒญํ‚น)
Chunk๋Š” ๋ฐ์ดํ„ฐ๋ฅผ ์ž‘๊ฒŒ ๋‚˜๋ˆ„๋Š” ๊ณผ์ •์ด๋‚˜ ๊ทธ ๋‹จ์œ„๋ฅผ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค. ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ(NLP)์—์„œ๋Š” ์ฃผ๋กœ ๊ธด ๋ฌธ์„œ๋ฅผ ์ž‘๊ฒŒ ๋‚˜๋ˆ„์–ด ์ฒ˜๋ฆฌํ•˜๋Š”๋ฐ ์‚ฌ์šฉ๋ฉ๋‹ˆ๋‹ค. ์˜ˆ๋ฅผ ๋“ค์–ด, ๊ธด ๋ฌธ์„œ๋‚˜ ์ฑ…์„ ์—ฌ๋Ÿฌ ๊ฐœ์˜ ์ฒญํฌ๋กœ ๋‚˜๋ˆ„์–ด ๊ฐ๊ฐ์˜ ์ฒญํฌ์—์„œ ์˜๋ฏธ๋ฅผ ์ถ”์ถœํ•œ ํ›„, ์ตœ์ข…์ ์œผ๋กœ ์ด๋ฅผ ํ†ตํ•ฉํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ์‚ฌ์šฉํ•ฉ๋‹ˆ๋‹ค. ์ฒญํ‚น์€ ๊ฒ€์ƒ‰ ํšจ์œจ์„ฑ์„ ๋†’์ด๊ณ , ์–ธ์–ด ๋ชจ๋ธ์ด ๋ณด๋‹ค ์งง์€ ๋ฌธ๋งฅ ๋‚ด์—์„œ ์ž‘์—…ํ•  ์ˆ˜ ์žˆ๋„๋ก ๋•์Šต๋‹ˆ๋‹ค.
์ฒญํฌ ๋‹จ์œ„๋Š” ์ ์ ˆํ•œ ํฌ๊ธฐ๋กœ ์„ค์ •ํ•˜๋Š” ๊ฒƒ์ด ์ค‘์š”ํ•ฉ๋‹ˆ๋‹ค. ๋„ˆ๋ฌด ์ž‘์œผ๋ฉด ์˜๋ฏธ๊ฐ€ ํ‡ด์ƒ‰๋˜๊ณ , ๋„ˆ๋ฌด ํฌ๋ฉด ๋ชจ๋ธ์ด ๋ฉ”๋ชจ๋ฆฌ ์ œ์•ฝ์— ๊ฑธ๋ฆด ์ˆ˜ ์žˆ๊ธฐ ๋•Œ๋ฌธ์ž…๋‹ˆ๋‹ค.
4. Vector DB (๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค)
Vector DB๋Š” ๋ฒกํ„ฐ(์ˆซ์ž๋กœ ํ‘œํ˜„๋œ ๋ฐ์ดํ„ฐ)๋ฅผ ์ €์žฅํ•˜๊ณ  ๊ฒ€์ƒ‰ํ•˜๋Š” ๋ฐ ํŠนํ™”๋œ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค์ž…๋‹ˆ๋‹ค. ์–ธ์–ด ๋ชจ๋ธ์ด๋‚˜ ์ด๋ฏธ์ง€ ๋ชจ๋ธ์—์„œ ํ…์ŠคํŠธ๋‚˜ ์ด๋ฏธ์ง€๋ฅผ ๋ฒกํ„ฐ ํ˜•ํƒœ๋กœ ๋ณ€ํ™˜ํ•œ ํ›„, ์ด๋ฅผ ๋น ๋ฅด๊ฒŒ ๊ฒ€์ƒ‰ํ•  ์ˆ˜ ์žˆ๋„๋ก ๋•์Šต๋‹ˆ๋‹ค.
์ฃผ๋กœ ์ž„๋ฒ ๋”ฉ(Embedding) ๊ณผ์ •์„ ํ†ตํ•ด ํ…์ŠคํŠธ๋‚˜ ์ด๋ฏธ์ง€๊ฐ€ ๋ฒกํ„ฐ๋กœ ๋ณ€ํ™˜๋˜๋ฉฐ, ์ด ๋ฒกํ„ฐ๋“ค์€ ๊ณ ์ฐจ์› ๊ณต๊ฐ„์— ์ €์žฅ๋ฉ๋‹ˆ๋‹ค. ๊ทธ ํ›„, ์‚ฌ์šฉ์ž๊ฐ€ ์ž…๋ ฅํ•œ ์ฟผ๋ฆฌ(์˜ˆ: ์งˆ๋ฌธ)๊ฐ€ ๋ฒกํ„ฐ๋กœ ๋ณ€ํ™˜๋˜์–ด ๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค์— ์ €์žฅ๋œ ๋‹ค๋ฅธ ๋ฒกํ„ฐ๋“ค๊ณผ ๋น„๊ต๋˜์–ด ๊ฐ€์žฅ ์œ ์‚ฌํ•œ ๊ฒฐ๊ณผ๋ฅผ ์ฐพ์•„๋ƒ…๋‹ˆ๋‹ค.
๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค๋Š” ์ฃผ๋กœ RAG ์‹œ์Šคํ…œ์—์„œ ์ •๋ณด ๊ฒ€์ƒ‰์„ ํšจ์œจ์ ์œผ๋กœ ์ˆ˜ํ–‰ํ•˜๋Š” ๋ฐ ํ•„์ˆ˜์ ์ž…๋‹ˆ๋‹ค. ๋Œ€ํ‘œ์ ์ธ ๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค๋กœ๋Š” Pinecone, FAISS(Facebook AI Similarity Search), Weaviate ๋“ฑ์ด ์žˆ์Šต๋‹ˆ๋‹ค.
์š”์•ฝ
LangChain: ์–ธ์–ด ๋ชจ๋ธ ํ™œ์šฉ์„ ์œ„ํ•œ ํ”„๋ ˆ์ž„์›Œํฌ.
RAG: ์ •๋ณด ๊ฒ€์ƒ‰๊ณผ ์–ธ์–ด ๋ชจ๋ธ ์ƒ์„ฑ์„ ๊ฒฐํ•ฉํ•œ ๊ธฐ๋ฒ•.
Chunk: ๋ฐ์ดํ„ฐ๋ฅผ ์ž‘๊ฒŒ ๋‚˜๋ˆ„์–ด ์ฒ˜๋ฆฌํ•˜๋Š” ๋‹จ์œ„.
Vector DB: ๋ฒกํ„ฐ ๋ฐ์ดํ„ฐ๋ฅผ ์ €์žฅํ•˜๊ณ  ๊ฒ€์ƒ‰ํ•˜๋Š” ํŠนํ™”๋œ ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค.
์ด ๊ธฐ์ˆ ๋“ค์€ ๋ชจ๋‘ ์ž์—ฐ์–ด ์ฒ˜๋ฆฌ์™€ ๋Œ€ํ˜• ์–ธ์–ด ๋ชจ๋ธ์˜ ํšจ์œจ์  ํ™œ์šฉ์„ ์œ„ํ•œ ํ•ต์‹ฌ์ ์ธ ์š”์†Œ์ž…๋‹ˆ๋‹ค.
0 notes
jbird-the-manwich ยท 3 months ago
Text
To answer your question, yes, they can and do search the internet (if asked, and if the specific bot supports it).
The llm itself, in most reasonable setups, is basically a parser for user intent. That's why it's not really that big of a deal that they guess the next token - that's the best thing they *could* do. They don't need to "know" things. They just need to be able to guess what the user means without expecting exact string literals, and be able to guess tokens to put together useable language, and synthesize data fed to them from other functions.
User asks a question, optionally telling the llm to search online. The llm outputs a function call requesting internet search of the inference code. Inference code catches this, runs a number of searches (anywhere from one to several tens, depending on the bot, the user, and the content), related data is sniffed from the search, usually by a smaller model, and passed back to the llm, who's job is then to summarize for the user. This isn't the only way they can reference data, but is in a sense a sort of web-mediated Retrieval Augmented Generation, which works the same way - documents are converted into a vector database for fast indexing of "what" and "where". User asks a question. Smaller model queries vector DB for relation to user input. If matches are found, relevant text is passed to the llm to summarise back to the user. This is one way that LLMs can be adapted to certain domains - by making domain specific data available to them. (and finetuning but thats in the weeds from here)
and on the topic of internet search and RAG, small local models can do this, as well, with plugins to search the internet, as can the models of most inference providers.
Though, depending on what the model has been trained on, it can sometimes have a useable knowledge for certain domains without access to the internet, but, in general, yes, the llm itself is a 3 dimensional array of floating point values that spits a response. a text engine. But it's only the language core. Which is adapted for different use cases by inference code. This is one reason LLMs and AI based on them are difficult to discourse about meaningfully, because we could be talking about the model (a set of frozen floating point values in memory) or its interface, or the functions made available to it, or the output of all of that, together, and most people only have the barest grasp of what the model even is, let alone to throw in the complexity of functions that may or may not be there depending on the software surrounding the model in the implementation.
tldr; yes they can google and how much they can google is alterable at inference time in code. The default for openrouter is five max searches per query but this can be changed by passing a parameter to the models api at inference time.
one of the things that really pisses me off about how companies are framing the narrative on text generators is that they've gone out of their way to establish that the primary thing they are For is to be asked questions, like factual questions, when this is in no sense what they're inherently good at and given how they work it's miraculous that it ever works at all.
They've even got people calling it a "ChatGPT search". Now, correct me if I'm wrong software mutuals, but as i understand it, no searching is actually happening, right? Not in the moment when you ask the question. Maybe this varies across interfaces; maybe the one they've got plugged into Google is in some sense responding to content fed to it in the moment out of a conventional web search, but your like chatbot interface LLM isn't searching shit is it, it's working off data it's already been trained on and it can only work off something that isn't in there if you feed it the new text
i would be far less annoyed if they were still pitching them as like virtual buddies you can talk to or short story generators or programs that can rephrase and edit text that you feed to them
#ai
76 notes ยท View notes
viperallc ยท 1 year ago
Text
How Alephium (ALPH) Revolutionizes Blockchain Technology
Tumblr media
Alephium is a cutting-edge sharded layer-one blockchain designed to overcome the limitations of existing blockchains, such as scalability, accessibility, and security. Itโ€™s an ideal platform for developers to create scalable decentralized applications (DApps) while offering individuals the benefits of decentralization and robust security.
Alephium focuses on solving todayโ€™s blockchain scalability and security issues by enhancing Proof-of-Work (PoW) and utilizing the Unspent Transaction Output (UTXO) model. Essentially, Alephium enables the creation of high-performance, accessible, and energy-efficient DApps and smart contracts.
Tumblr media
How Alephium Works
Alephium employs several innovative technologies to address the traditional blockchain drawbacks and improve scalability, programmability, security, and energy efficiency. Letโ€™s dive into these features.
Enhancing Scalability with BlockFlow Sharding
Alephium utilizes a sharding algorithm called BlockFlow to boost scalability. Sharding splits data into smaller, manageable parts called shards, facilitating parallel transactions. The UTXO model and Directed Acyclic Graph (DAG) data structure further aid effective sharding, allowing Alephium to handle around 10,000 transactions per second.
Boosting Energy Efficiency with Proof-of-Less-Work (PoLW)
The blockchain employs a unique Proof-of-Less-Work (PoLW) consensus mechanism, adjusting mining difficulty based on real-time network conditions. This approach significantly reduces energy consumption compared to traditional PoW algorithms.
Enhancing Programmability and Security with the UTXO Model
Alephium uses the UTXO model to enhance programmability and security, ensuring fast, efficient transactions. This model maintains the same level of security as Bitcoin while offering better scalability and flexibility.
Leveraging a Custom Virtual Machine and Programming Language
Alephium has its own virtual machine, SDK, and a performance-optimized programming language. These tools include built-in security features that prevent unauthorized transactions and common attack vectors. Developers can leverage these innovations to build advanced DApps and smart contracts.
What Makes Alephium Unique?
Alephium stands out from other blockchains with its unique combination of features designed to improve scalability, security, and energy efficiency.
Maximizing Efficiency with Sharding
Sharding divides the network into smaller, manageable subsets called shards, each acting as an independent blockchain. This allows for parallel transaction processing, distributing the workload across multiple shards and increasing overall throughput and network capacity.
Leveraging the UTXO Model for Enhanced Security and Flexibility
The UTXO model uses unspent transaction outputs as inputs for new transactions, enhancing scalability and programmability. This model ensures secure and efficient transactions while maintaining Bitcoin-level security.
Achieving Energy Efficiency with Proof-of-Less-Work (PoLW)
Alephiumโ€™s PoLW consensus mechanism minimizes energy consumption compared to traditional PoW algorithms. This makes Alephium much more energy-efficient than Bitcoin.
Custom Virtual Machine for Superior Performance
Alephiumโ€™s custom VM, Alphred, addresses the drawbacks of existing DApp platforms by improving security, scalability, and programmability. It enables developers to create Peer-to-Peer (P2P) smart contracts with ease.
Ralph: A Unique Programming Language for DApps
Alephium features its own programming language, Ralph, specifically designed for building secure and efficient DApps and smart contracts. This empowers businesses and individuals to leverage Alephiumโ€™s robust blockchain platform.
Tumblr media
โœ“ Manufacturer:ย Bitmain โœ“ Model:ย Antminer AL3 โœ“ Supported Algorithm:ย Alephium (ALPH) โœ“ Hashrate:ย 8 TH/s โœ“ Power Consumption:ย 3200W โœ“ Dimensions:ย 195 x 290 x 430 mm โœ“ Weight:ย 14.2 kg โœ“ Operating Noise Level:ย 75 dB โœ“ Power Supply Unit:ย Included โœ“ Release Date:ย August 2024 โœ“ Warranty:ย 1 year manufacturer repair or replace
Wrapping Up
Alephium provides a scalable and secure blockchain platform with innovative features like sharding, the UTXO model, and PoLW consensus. These elements make Alephium a powerful tool for developers and individuals looking to create reliable and efficient decentralized applications.
Muhammad Hussnain Visit us on social media:ย Facebookย |ย Twitterย |ย LinkedInย |ย Instagramย |ย YouTubeย TikTok
0 notes