I think there are a lot of technological advancements that are easy to "not like" when you know some select few details about them. Including literal sausage making as another commenter on here mentioned.
T-shirts are great until you hear about the conditions of where they're made and how their disposal is managed. Social media is great until you realize how much they know about you and how they use that knowledge. Modern medicine is easy to not like when you look at the animal experiments that made it happen. And again, sausages - I know some vegetarian folks who are vegetarian in protest of how most meat is produced.
I kind of wonder if there is a subset of comfortable modern society where every aspect is easily likable no matter how much you know about it. Bonus points if that society is environmentally sustainable.
Sure, but I think there is a key diff between those examples and LLMs. In those cases, people don't necessarily mind the machines involved in the process, they dislike the socioeconomic and labor structures around those machines, or the animal cruelty—this is a side effect of the social organization of labor, not of the machine itself. Here, contrarily, the claim is that the machine itself is the thing that becomes less impressive once you know how it works-this is also likely true of any machine, but the point is not about the demystification of the technical function in the tshirt and sausage examples.
Life is [generally] inherently exploitative. Vegetarianism and veganism is a first world luxury, the bottom line is animal protein consumption is a readily available high-quality caloric intake option to encourage children to grow into healthy adults*.
Learning to exist in the world and hold the uncomfortable parts of being a human being has been a valuable and useful skill in my life across many dimensions.
* I'm not advocating for eating excessive meat, but going to extremes to avoid it does suggests something other factor is in play.
> Vegetarianism and veganism is a first world luxury
"Luxury" makes me choke : animal protein has always been seen as a luxury meal in most cultures, one that you don’t have every day, as found in any authors from XIX or early XX and you’ll never seen animal protein as a poor food, let alone meat [0]
For 1kg of plant protein you’ll get 230gr of milk protein, 220gr egg’s prot 150gr chicken prot, 120gr porc prot or 70gr beef prot. There’s some difference in proteins availability but in a way smaller scale. Regarding the quality : most traditional societies have combined plant proteins in their daily diets : corn and beans (Latin American), cheakpeas and bulgur (Middle East) rice and lentils (India), soyfood and rice (China, Japan, Korea), millet and peanuts (Africa), rice and tempeh (Java, a very poor place with one the biggest density of the world). The list can go long. Modern science showed us those associations skyrocket the ratio of limiting amino acids, like Lysin + Cystine/Methionine, for soy and brown rice.
Plants have always been the poor’s staple protein because they are dirt cheap, nutritious and convenient. More importantly for the future: the require many times less land to grow because of the protein consumed/returned by animals ratio (see above). Wild fish is an exception but the relative yields have been decreasing for decades, only compensated by more and bigger boats. Also 1/5 of world wild fishing production exist for feeding livestock and farmed fishes.
It’s true (and sad in my opinion) plant proteins are now luxury in certains cities/places, the reasons for that are habits and taste preferences, world richness imbalance (Europe imports 60kg/person/year of Brazilian soy only for livestock consumption, a quantity that would provide 50% of their protein need if consumed directly as so) and subsidies and price regulations to maintain an unsustainable diary industry.
I’m not advocating for everyone eating soy but let’s face it : feeding livestock with human consumable proteins is not the most efficient diet.
0 L’assomoir from Emile Zola, loosely translated by myself : "On Sundays, when there was work at the mine, we'd have a bit of bacon, and on the rare good days, a chunk of beef the size of your fist. It was a real treat"
If the average student is a bully, disruptive, and disrespectful to teachers then I think I might actually opt out of being around average people if possible.
I tried this (though with a different tool called aichat) for extremely simple stuff like just "convert this mov to mp4" and it generated overly complex commands that failed due to missing libraries. When I removed the "crap" from the commands, they worked.
So much like code assistance, they still need a fair amount of baby sitting. A good boost for experienced operators but might suck for beginners.
Plus you need to know the format of your source file to design the command correctly. How many audio tracks, is the first video track a thumbnail or the video, are the subtitles tracks forced, etc.
And in some situations ffmpeg has some warts you have to go around. Like they introduced recently a moronic change of behaviour where the first sub tracks becomes forced/default irrespective of the original forced/default flag of the source. You need to add "-default_mode infer_no_subs" to counter that.
To me it's more like a guy getting his first car and complaining that the car is driving him in a direction that may or may not be correct, despite his best efforts to steer it where he wants to go. And the only way to know whether he ends up in the right place is to get out of the car, look around, and maybe ask more experienced drivers. Failing that, his only option is to get back in and hope to be luckier in the next trip.
Or he can just ditch the car and walk. Sure, it's slower and requires more effort, but he knows exactly how to do that and where it will take him.
The beer brewers in my home town used to have a self-driving horse and cart which knew the daily delivery route going by all pubs and didn't really need a human to steer it or indeed be conscious during the trip. Expectedly, the delivery guy would get drunk first thing in the morning and just get carted about collecting the money.
Pony & trap could be largely self-driving, after an initial training period. That would have been a distinct negative to "upgrading" for some, I'd imagine.
I no longer check with these AI tools after a number of attempts. Unrelated, a friend thought there was a NFL football game last Saturday at noon. Checking with Google's Gemini, it said "no", but there was one between two teams whose season had ended two weeks before at 1:00 Eastern Time and 2:00 Central. (The times are backwards.)
libx264 is the best encoder for h264 ffmpeg has to offer so it's pretty important you bundle it in your ffmpeg install. Those commands are perfectly standard, I've been using something like that for 10+ years
I don't disagree that we need to be cautious with LLMs, but I've personally stopped asking GPT-4/GPT-4 mini for technical answers. Sonnet 3.5 and DeepSeek V3 (which is much cheaper but still not as good as Sonnet) are your best bet for technical questions.
Where I find GPT to perform better than Sonnet is with text processing. GPT seems to better understand what I want when it comes to processing documents.
I'm convinced that no LLM provider has created or will create a moat, and that we will always need to shop around for an answer.
Unfortunately you need to be tier 2 to use o1-mini. The only time I really use GPT is to summarize documents and for that, GPT-4o mini works well enough and it is significantly cheaper than other high quality models, so I never really rack up an OpenAI bill.
o1 is such a joke, worse than 4o in some ways like multiturn,
The months old sonnet feels a generation ahead of any OAI product I've used, I'll believe the hype on o3 when I see it, remember the sora and voice roll out?
I had this bizarre bug in rust networking code where packets were getting dropped.
i dumped all 20k lines into o1pro. it thought for about ten minutes and came back telling me that my packets had a chance of being merged if set in quick succession and i needed to send the length before each message and scan packets in a loop for subdivisions on the client. this bug hadnt happened before, only when running locally on a newer faster machine, and was frequent but hard to replicate.
it was correct, and provided detailed pseudo code to solve it.
the second case involved some front end code where during an auth flow ios would force refresh on returning to the browser causing authentication state to be lost. o1pro thought for about 5 minutes before telling me ios has a heuristic with which it decides to close an app on context switch based on available ram, etc, and that i needed to conditionally check for ios and store partial state in local store on leave assuming the app could be deloaded without my control.
it was correct. with some more back and forth we fixed the bug.
these are not the kinds of problems that claude and gpt<4 have been able to help with at all.
I also used voice, and video voice extensively for translation tasks in korea, japan, and taiwan, and for controlling japanese interfaces and forms for tax documents and software.
o1 is not a general-purpose model, and it's not very good at multi-turn; it should instead be given all the context upfront: https://www.latent.space/p/o1-skill-issue
what exactly do you want the llm to do here? if the ask was so unambiguous and simple that it could be reliably generated, then the interface wouldn't be so complicated to use in the first place! LLMs are not in any way best suited for one-shot prompt => perfect output, and expectations to that effect are extremely unreasonable. the reason why LLMs are still hard for beginners to use is because the software is hard to use correctly. as with LLM output goes life itself: the results you get from using a tool can only ever be as good as the (mental) model used to choose that tool & the inputs to begin with. if all the information required to generate the output were contained by the initial prompt, then there would be absolutely no need to use the LLM at all in the first place.
Hate to be that guy, but which LLM was doing the generation? GPT-4 Turbo / Claude 3.x have not really let me down in generating ffmpeg commands - especially for basic requests - with most of their failures resulting from ___domain-specific vagaries that an expert would need to weigh in on m
You have a fair point. Some LLMs are better at some tasks, and prompts can make a difference no doubt.
Perhaps at some point there will be a triage LLM to slurp up the problem and then decide which secondary LLM is most optimal for that query, and some tertiary LLMs that execute and evaluate it in a virtual machine, etc.
Oh I talked to some guys who started a company that does that. This was at an AI meetup in SF last year. They were mainly focused on making $/token cheaper by directing easy/dumb queries to smaller dumber models, but it also increases output quality because some models are just better at certain things. I'm sure all the big companies already have implementations of this by now even if they don't use it everywhere
"Writing working commands first try for every single ffmpeg feature that exists" is the highest bar I've ever heard of, I love it. I'm gonna start listing it as a requirement on job postings. Like an ffmpeg speedrun.
I don't think there's a single human on or outside of this planet that can meet that requirement, but Claude has been pretty good to me. It's certainly a much better starting point than pouring over docs and SO posts.
I know I struggled on getting a good command to “simply” make the videos from my Z8 smaller (in file size).
Usually the color was wrong and I don’t care enough to learn about colorspaces to figure out how to fix it and it’s utterly insane how difficult it is even with LLMs.
Just reencode it as is but a little more lossy. Is that so hard?
I think in the non LLM world though you at least have the trail of documentation you can unwind once you're in a bind. I don't care for prompt-a-mole fighting.
Awesome idea. At first, the LLM was spitting out live links to sometimes-real websites, which I thought was a bit too weird so I told it to not link to external sites. But linking to "internal" sites sounds fun actually....
DBus is just an RPC layer. You still need to define a standard DBus interface for any given feature. Not all UNIX-likes ship DBus in the base system (which is certainly one component of why Sway has Swaysock.)
Meanwhile, every Wayland compositor speaks at least the core Wayland protocols and probably some of the extensions, and they all go through the same standardization process, whereas there's certainly no such process for DBus or random UNIX ___domain sockets based protocols. It's simpler to just use Wayland protocols where possible.