This account is being kept for the posterity, but it won’t see further activity past February.

If you want to contact me, I’m at /u/lvxferre@mander.xyz

  • 7 Posts
  • 1.06K Comments
Joined 4 years ago
cake
Cake day: April 9th, 2021

help-circle




  • Yup. Google consistently gets rid of features or services that it deems unprofitable. And that’s fine, really - as long as you don’t pretend that you’re doing it for the users.

    To be fair in modern phones there are some features that if removed would make the user experience better.

    I hear ya - for example, the SIM toolkit being able to send you pop-ups (phone providers use that to spam the users).


  • We’re removing some underutilized features in Google Assistant to focus on delivering the best possible user experience.

    Is this the non sequitur used nowadays to explain removal of features? “We’re removing it to give you a better experience”??? That’s bloody hilarious.

    Be honest at least dammit. If you don’t want to maintain a feature, because it’s against your best interests, say so. Users are not stupid, and should not be implied to be stupid with this idiotic “it’s for you lol” discourse.

    (I don’t even use Botnet Assistant.)



  • I agree too much with the text to comment anything meaningful about it. So let’s see the comments…

    One aspect of the spread of LLMs is that we have lost a useful heuristic. Poor spelling and grammar used to be a signal used to quickly filter out worthless posts. […]

    Although I agree with the title, I also don’t think the internet is that significantly different from before GPTs 4, 3, or 2. Articles written by interns or Indian virtual assistants about generic topics are pretty much as bad as most AI generated material […]

    Both comments reminded me a blogpost that I wrote more than a year ago, regarding chatGPT-3. It still applies rather well to 2024 LLMs, and it shows what those two tech bros are missing, so I’ll copypaste it here.

    ###The problem with GPT3.

    Consider the following two examples.

    Example A.

    GPT3 bots trained on the arsehole of the internet (Reddit), chatting among themselves:

    1. LPT: When you’re planning your next vacation, be sure to consider the cost of airfare and hotels. You may not have the money to go on a luxury trip, but at least you won’t have to worry about the economy when you leave! // You’re worth it!
    2. (to #1) This is great advice! I definitely need to start considering the cost of airfare and hotels when planning my next vacation. Thank you for the tips!
    3. (to #2) I completely agree! When I’m planning my next vacation, my number one priority is to make sure that it’s worth the cost. Thanks for the advice!
    4. (to #2) Keep up the great work!

    The grammar is fine, and yet those messages don’t say jack shit.

    • #1 - the so-called “life pro tip” is fairly obvious, since people are aware that airfare and hotels cost quite a bit. It is not informing the reader about something potentially missed.
    • #1 - “You may not have the money to go on a luxury trip” is extraneous, no matter how you interpret it, as it’s directed towards people who won’t fly and spend time in hotels.
    • #1 - How the hell are you expected to worry less or more about the economy, depending on how you plan your vacations?
    • #1 - you’re worth… what? The vacations? Not worrying about the economy? Something else?
    • #2 - needlessly repeating a huge chunk of #1.
    • #3 and #4 - it’s clear that #1 and #2 are different participants, #2 provided nothing worth thanking, and yet it’s still being thanked. Why?

    Example B.

    Human translation made by someone with not-so-good grasp of the target language.

    Captain: What happen ?
    Mechanic: Somebody set up us the bomb.
    Operator: We get signal.
    Captain: What !
    Operator: Main screen turn on.
    Captain: It's you !!
    CATS: How are you gentlemen !!
    CATS: All your base are belong to us.
    CATS: You are on the way to destruction.
    

    The grammar is so broken that this excerpt became a meme. And yet you can still retrieve meaning from it:

    • Captain, Mechanic and Operator are the crew of a ship
    • Captain asks for info
    • Someone is trying to kill them with a bomb
    • Operator and Mechanic inform Captain on what happens
    • CATS sarcastically greets the crew, and provides them info to make them feel hopeless
    • Captain expresses distress towards CATS

    What’s the difference? It’s purpose. In (B) we can give each utterance a purpose, even if the characters are fictional - because they were written by a human being. However, we cannot do the same in (A), because the current AI-generated text does not model that purpose.

    And yes, assigning purpose to your utterances is part of the language. Not just what tech bros are able to see, namely: syntax, morphology, and spelling.


  • Archive link.

    Personal take: suck it up, Somalia; if the population of Somaliland has effective control of the region, and desires it to be independent, then there isn’t much that you could (or should) do. And from that, if both Somaliland and Ethiopia reach an amicable agreement over the ports, so be it.

    Also, let us drop all that babble about territorial integrity. Even if you believe in this sort of political superstition, Somalia’s territorial integrity went kaboom in 1991.









  • Glacial = anhydrous. People call it this way because pure acetic acid has a rather high freezing point (16°C), and it looks a lot like plain ice when frozen. (It still stinks vinegar once you open the bottle though.) But once you add even a bit of water, the freezing point drops considerably, so acetic acid solutions don’t show the same “ice”.

    So in colder days, you need to rewarm it back into a liquid. Then people get really sloppy (I know it not just from that professor’s anecdote, but from watching it). They say “I’m just rewarming it, and it’s just acetic acid, what could go wrong?”. Well, it’s still a big flask of a corrosive, volatile, and flammable substance.

    In the meantime, the same people doing dangerous reactions like nitration (it literally explodes if you let it get too hot - spreading nitric acid, sulphuric acid, and some carcinogenic solvent) “miraculously” pay full attention, obsessively taking care of the temperature of the ice bath.

    Part of the advice that I mentioned in that comment chain is that - smaller dangers are still dangers, do not underestimate them.



  • Most “rules of thumb” become awful advice when used indiscriminately.

    People assign slightly different meanings to the same words. You need to acknowledge this to understand what they say.

    Words also change meaning depending on the context.

    When you still don’t get what someone else said, it’s often more useful to think that you’re lacking a key piece of info than to assume that the other person does.

    Hell is paved with good intentions. This piece of advice is popular, but still not heard enough.

    Related to the above: if someone in your life is consistently rushing towards conclusions, based on little to no information, minimise the impact of that person in your life.

    Have at least one recipe using leftovers of other recipes. It’ll reduce waste.

    Alcohol vinegar is bland, boring, and awful for cooking. But it’s a great cleaning agent.

    Identify what you need to keep vs. throw away. Don’t “default” this indiscriminately, analyse it on a per case basis.

    The world does not revolve around your belly button and nature won’t “magically” change because of your feelings.

    You can cultivate herbs in a backyard. No backyard? Flower pots. No flower pots? Old margarine pot. (Check which herbs grow well where you live.)