*these are not my banners, I got them from someone else. Feel free to use them tho*
(guy who hasnt ever seen anything) i dont know what that is. it kind of reminds me of nothing
A fun meme for y'all
i’m sorry what
Pog as slang comes from the twitch emote, taken from a video with pogs in them. What is a pog? A small cap from a juice drink that is traded and used in games by 90s kids. The juice drink was called POG, short for passion fruit orange guava. Passion fruit was named after its flower, the passion flower. Passion flowers were called "flower of the five wounds" in Portuguese, eventually being translated as passion fruit by the English in reference to the passion of Jesus, aka his last moments alive before being crucified.
I love history, language, words, communication, and humans.
You gotta be careful with tumblr girls because you'll flirt a little and then all of the sudden they're curled up inside your ribcage.
Honestly I'm pretty tired of supporting nostalgebraist-autoresponder. Going to wind down the project some time before the end of this year.
Posting this mainly to get the idea out there, I guess.
This project has taken an immense amount of effort from me over the years, and still does, even when it's just in maintenance mode.
Today some mysterious system update (or something) made the model no longer fit on the GPU I normally use for it, despite all the same code and settings on my end.
This exact kind of thing happened once before this year, and I eventually figured it out, but I haven't figured this one out yet. This problem consumed several hours of what was meant to be a relaxing Sunday. Based on past experience, getting to the bottom of the issue would take many more hours.
My options in the short term are to
A. spend (even) more money per unit time, by renting a more powerful GPU to do the same damn thing I know the less powerful one can do (it was doing it this morning!), or
B. silently reduce the context window length by a large amount (and thus the "smartness" of the output, to some degree) to allow the model to fit on the old GPU.
Things like this happen all the time, behind the scenes.
I don't want to be doing this for another year, much less several years. I don't want to be doing it at all.
----
In 2019 and 2020, it was fun to make a GPT-2 autoresponder bot.
Hardly anyone else was doing anything like it. I wasn't the most qualified person in the world to do it, and I didn't do the best possible job, but who cares? I learned a lot, and the really competent tech bros of 2019 were off doing something else.
And it was fun to watch the bot "pretend to be me" while interacting (mostly) with my actual group of tumblr mutuals.
In 2023, everyone and their grandmother is making some kind of "gen AI" app. They are helped along by a dizzying array of tools, cranked out by hyper-competent tech bros with apparently infinite reserves of free time.
There are so many of these tools and demos. Every week it seems like there are a hundred more; it feels like every day I wake up and am expected to be familiar with a hundred more vaguely nostalgebraist-autoresponder-shaped things.
And every one of them is vastly better-engineered than my own hacky efforts. They build on each other, and reap the accelerating returns.
I've tended to do everything first, ahead of the curve, in my own way. This is what I like doing. Going out into unexplored wilderness, not really knowing what I'm doing, without any maps.
Later, hundreds of others with go to the same place. They'll make maps, and share them. They'll go there again and again, learning to make the expeditions systematically. They'll make an optimized industrial process of it. Meanwhile, I'll be locked in to my own cottage-industry mode of production.
Being the first to do something means you end up eventually being the worst.
----
I had a GPT chatbot in 2019, before GPT-3 existed. I don't think Huggingface Transformers existed, either. I used the primitive tools that were available at the time, and built on them in my own way. These days, it is almost trivial to do the things I did, much better, with standardized tools.
I had a denoising diffusion image generator in 2021, before DALLE-2 or Stable Diffusion or Huggingface Diffusers. I used the primitive tools that were available at the time, and built on them in my own way. These days, it is almost trivial to do the things I did, much better, with standardized tools.
Earlier this year, I was (probably) one the first people to finetune LLaMA. I manually strapped LoRA and 8-bit quantization onto the original codebase, figuring out everything the hard way. It was fun.
Just a few months later, and your grandmother is probably running LLaMA on her toaster as we speak. My homegrown methods look hopelessly antiquated. I think everyone's doing 4-bit quantization now?
(Are they? I can't keep track anymore -- the hyper-competent tech bros are too damn fast. A few months from now the thing will be probably be quantized to -1 bits, somehow. It'll be running in your phone's browser. And it'll be using RLHF, except no, it'll be using some successor to RLHF that everyone's hyping up at the time...)
"You have a GPT chatbot?" someone will ask me. "I assume you're using AutoLangGPTLayerPrompt?"
No, no, I'm not. I'm trying to debug obscure CUDA issues on a Sunday so my bot can carry on talking to a thousand strangers, every one of whom is asking it something like "PENIS PENIS PENIS."
Only I am capable of unplugging the blockage and giving the "PENIS PENIS PENIS" askers the responses they crave. ("Which is ... what, exactly?", one might justly wonder.) No one else would fully understand the nature of the bug. It is special to my own bizarre, antiquated, homegrown system.
I must have one of the longest-running GPT chatbots in existence, by now. Possibly the longest-running one?
I like doing new things. I like hacking through uncharted wilderness. The world of GPT chatbots has long since ceased to provide this kind of value to me.
I want to cede this ground to the LLaMA techbros and the prompt engineers. It is not my wilderness anymore.
I miss wilderness. Maybe I will find a new patch of it, in some new place, that no one cares about yet.
----
Even in 2023, there isn't really anything else out there quite like Frank. But there could be.
If you want to develop some sort of Frank-like thing, there has never been a better time than now. Everyone and their grandmother is doing it.
"But -- but how, exactly?"
Don't ask me. I don't know. This isn't my area anymore.
There has never been a better time to make a GPT chatbot -- for everyone except me, that is.
Ask the techbros, the prompt engineers, the grandmas running OpenChatGPT on their ironing boards. They are doing what I did, faster and easier and better, in their sleep. Ask them.
You can't spell advertisements without putting semen between tits
platonic cuddling. platonic messing with your hair. platonic kissing
Round 1, Week 2: Match 18 of 32
Propaganda as submitted by YOU, in initial entries, tags, and reblogs:
Schmigadoon:
- "Schmigadoon is one of the funniest show I've seen ever! it's a musical in which the two main characters ARE NOT from the musical universe ... the songs are super catchy and the plot is so sweet. it's 6 50 minute episodes I swear it's worth watching"
- do it for aaron tveit
Game Changer:
- "The best game show"
IT'S A GAME CHANGER!!!!
Pierre Balmain Haute Couture, 1953
💕💕💕💕💕💕💕
Tempted to think of this merely as another stage of "What Fresh Hell Is This."
The glass cliff, on the other hand, refers to the phenomenon by which women are more likely to be appointed to senior executive positions during times of organizational crisis, making them less likely to succeed. These newly appointed executives may confront internal board resistance, operate with less time flexibility, and ultimately receive shorter tenure than their male counterparts. And, when a woman CEO is terminated from her position, she is more likely than not to be replaced by a male (the “savior effect”).
He's stepping down so people don't pay attention to Tesla recalling 1 mil. cars.
They’re WHAT???
Also the woman named is an alt-right Trump supporter so....
It’s the second large recall this year! 🤗 The first was 350k cars in the US in February. The year isn’t even half over.




















