How might this tool work in terms of “archiving” a site? This is just something I was wondering given the recent change and controversy about archiving service sites on Wikipedia.
Site Spy keeps snapshot history, so you can revisit older versions of a page and inspect how it changed over time, not just get the latest alert. I’d describe it more as monitoring with retained history than as a dedicated public archive, but deeper archival integrations are definitely something I’ve thought about
I had a couple of experiences where I suspected I was hearing LLM-generated/edited text being read aloud. It was at two different webinars about that were about roadmaps or case studies about some products that I use. It was a bit uncanny because I could detect the stylistic patterns ("It's not X, it's Y" and "No X, no Y, just Z"), but it was kind of jarring to see them spoken by a person on a video call. It makes me think this kind of pattern might be engaging, but for a lot of people, it now sticks out for the wrong reasons.
Once LLM generated speech or content start getting into the live answers of Q&A sessions, that would be sad. I know some people try to get through interviews, but I think that might be a bit harder to not detect.
Floaters are very annoying. One solution for me has been to turn my brightness down as much as possible, but of course this might make it difficult to see what's on your monitor if there's too much ambient light around you.
This reminds me of a news story from many years ago about a research study that won the Ig Nobel Prize. The research study was about how almost all mammals weighing over 3 kg take just about the same amount of time to empty their bladders when urinating.
I need to know if this is also true of aquatic animals. A blue whale is ~50x the size of an African elephant. Does that mean that a blue whale expels some 2100 gallons (~8000 liters) in 20 seconds?
Right. I think when these appear in some documentation related to computing, they should also mention whether it is using these words in compliance with RFC 2119 or RFC 6919.
In a way this kind of reminds me of how in some religions or cultures, they may try to warn you away from using Oujia boards or Tarot, or really anything where you are doing divination. I suppose because in a way, it could lead to an uncharted exploration of heavy topics.
I’m not a heavy user of LLMs and I’m not sure how delusional I could be, but I wonder if a lot of these things could be prevented if people could only send like one or two follow up messages per conversation, and if the LLM’s memory was turned off. But then I suppose this would be really bad for the AI companies’ metrics. Not sure how it would impact healthy users’ productivity either. Any thoughts?
Not just the metrics, the actual utility. For the things the LLMs are good at, the context matters a lot; it's one of the things that makes them more than glorified ELIZA chatbots or simple Markov chains. To give a concrete example: LLMs underpin the code editing tools in things like Copilot. And all that context is key to allow the tool to "reason" through the structure of a codebase.
But they should probably come with a big warning label that says something to the effect of "IF YOU TALK ABOUT YOURSELF, THE NATURE OF THE MACHINE IS THAT IT WILL COME TO AGREE WITH WHAT YOU SAY."
After I read this, I started to look at the Wikipedia article on Base64 and eventually got to the article for the data URI scheme. That's where I found a sentence that seems to a little bit at odds with the blogpost. The Wikipedia article mentions that "whitespace characters are not permitted in data URIs".
But then I suppose it goes back to the main thrust of the blogpost because it says that in the context of HTML 4 and 5, that linefeeds within an attribute value are ignored. So possibly there are some other contexts where whitespace might not be ignored.
They are not, but you can encode them, if you encode whitespace characters, you included whitespace in a URL.
One of the requirement of URLs is that it needs to be transmissible over paper or aural media, so arbitrary octets and the unused portion of ASCII are not legal either.
I disagree, because there’s always a chunk of advertising that seems to be all about targeting low-income or people who aren’t financially savvy and I don’t think it’s ethical for an apparatus to take advantage of them.
What utility does a box of cookies have? A bar of chocolate? A can of soda? Those things are about pleasure and have serious harmful consequences if overused - just like tobacco, alcohol and drugs.
What about video games? They only have utility in pleasure and the sedentary lifestyle associated with over-playing them is extremely harmful.
Sounds to me like you have some random things you decided you don't like and want to ban ads for them, not that you've done any thinking about utility (other than as a bad attempt at rationalizing your anti-some things campaign).
reply