

You should consider talking to a therapist about why you don’t consider non-romantic relationships to be valuable, because your take is not only not universal, but also indicative that you have some shit you need to work through.


You should consider talking to a therapist about why you don’t consider non-romantic relationships to be valuable, because your take is not only not universal, but also indicative that you have some shit you need to work through.


IMO, yes. I think it would make people more, rather than less, inclined to comment on a cross-post made in a smaller communities, since then their comment would be more visible.
The main concern I can see being raised is potentially leading to brigading? I’m not sure if that’s much of an issue on Lemmy and I would assume being able to de-federate would mitigate that substantially.
I think that makes you “the guy who really likes to talk about Linux.”


Ars points out that these findings contradict those of other experiments and then goes on to postulate as to why. I clicked on the link to the other experiment:
when data is combined across three experiments and 4,867 developers, our analysis reveals a 26.08% increase (SE: 10.3%) in completed tasks among developers using the AI tool
By comparison, this experiment considered 16 developers. That’s 0.3% as many as the experiments its findings contradict. Fortunately, the authors don’t claim their findings are broadly applicable. They even have a table that reads:
We do not provide evidence that | Clarification —- | —- AI systems do not currently speed up many or most software developers | We do not claim that our developers or repositories represent a majority or plurality of software development work AI systems do not speed up individuals or groups in domains other than software de- velopment | We only study software development AI systems in the near future will not speed up developers in our exact setting | Progress is difficult to predict, and there has been substantial AI progress over the past five years [2] There are not ways of using existing AI systems more effectively to achieve positive speedup in our exact setting | Cursor does not sample many tokens from LLMs, it may not use optimal prompting/scaffolding, and domain/repository-specific training/finetuning/few-shot learning could yield positive speedup
That said, the study has been an interesting read so far. I highly recommend reading it directly rather than just the news posts about it. Check out their own blog post: https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/
I personally find the psychological effect - the devs thought they were 20% faster even afterward - to be pretty interesting, as it suggests that even if more time overall is spent, use of AI could reduce cognitive load and potentially side effects like burnout.
I’d like to see much larger scale studies set up like this, as well as studies of other real world situations. For example, how does this affect the amount of time this takes 10,000 different developers to onboard onto an unfamiliar repository?
It may be aware of them, but not in that context. If you asked it how to solve the problem rather than to solve the problem for you, there’s a chance it would suggest you use a reverse image search.
LLM image processing doesn’t work the same way reverse image lookup does.
Tldr explanation: Multimodal LLMs turn pictures into a thousand 200-500 or so words tokens, but reverse image lookups create perceptual hashes of images and look the hash of your uploaded image up in a database.
Much longer explanation:
Multimodal LLMs (technically, LMMs - large multimodal models) use vision transformers to turn images into tokens. They use tokens for words, too, but these tokens don’t also correspond to words. There are multiple ways this could be implemented, but a common approach is to break the image down into a grid, then transform each “patch” of a specific size, e.g., 16x16, into a single token. The patches aren’t transformed individually - the whole image is processed together, in context - but it still comes out of it with basically 200 or so tokens that allow it to respond to the image, the same way it would respond to text.
Current vision transformers also struggle with spatial awareness. They embed basic positional data into the tokens but it’s fragile and unsophisticated when it comes to spatial awareness. Fortunately there’s a lot to explore in that area so I’m sure there will continue to be improvements.
One example improvement, beyond improved spatial embeddings, would be to use a dynamic vision transformers that’s dependent on the context, or that can re-evaluate an image based off new information. Outside the use of vision transformers, simply training LMMs to use other tools on images when appropriate can potentially help with many of LMM image processing’s current shortcomings.
Given all that, asking an LLM to find the album for you is like - assuming you’ve given it the ability and permission to search the web - like showing the image to someone with no context, then them to help you find what music video - that they’ve never seen, by an artist whose appearance they describe with 10-20 generic words, none of which are their name - it’s in, and to hope there were, and that they remembered, the specific details that would make it would come up in the top ten results if searched for on Google. That’s a convoluted way to say that it’s a hard task.
By contrast, reverse image lookup basically uses a perceptual hash generated for each image. It’s the tool that should be used for your particular problem, because it’s well suited for it. LLMs were the hammer and this problem was a torx screw.
Suggesting you use - or better, using a reverse image lookup tool itself - is what the LLM should do in this instance. But it would need to have been trained to think to suggest this, capable of using a tool that could do the lookup, and have both access and permission to do the lookup.
Here’s a paper that might help understand the gaps between LMMs and tasks built for that specific purpose: https://arxiv.org/html/2305.07895v7


PSTN is wiretapped.
It’s a good thing that the website itself supports sending and receiving alerts, then.


When did Democrats have a supermajority in both the House and the Senate?


I thought Hue bulbs used Zigbee?


The up arrow moves through the letters, e.g., A->B->C. The down arrow moves to the next character in the sequence, e.g., C->CA->CAA. If you click past the correct letter, you’ll have to click all the way through again. And if you submit the wrong letter, you have to start all over (after it takes twenty seconds attempting to connect with the wrong password and then alerts you that it didn’t work, of course).


Fair point, I should have asked about commercial games in general
That said I didn’t mean that the game studio itself would do the AI training and own their models in-house; if they did, I’d expect it to go just as poorly as you would. Rather, I’d expect the model to be created by an organization specialized in that sort of thing.
For example, “Marey” is one example I found of a GenAI model that its creators are saying was trained ethically.
Another is Adobe Firefly, where Adobe says they trained only on licensed and public domain content. It also sounds like Adobe is paying the artists whose content was used for AI training. I believe that Canva is doing something similar.
StabilityAI is also doing something similar with Stable Audio 2.0, where they partnered with a music licensing company, AudioSparx, to ensure that artists are compensated, AI opt outs are respected, etc…
I haven’t dug into any of those too deep, but they seem to be heading in the right direction at the surface level, at least.
One of the GenAI scenarios that’s the most terrifying to me is the idea of a company like Disney using all the material they have copyright for to train their own, proprietary GenAI image, audio, and video tools… not because I think the outputs would be bad, but because of the impact that would have on creators in that industry.
Fortunately, as long as copyright doesn’t apply to purely AI generated outputs, even if trained entirely on your own content, then I don’t think Disney specifically will do this.
I mention that as an example because that usage of AI, regardless of how ethically the model was trained, would still be unethical, in my opinion. Likewise in game creation, an ethically trained and operated model could still be used unethically to eliminate many people’s jobs in the interest solely of better profits.
I’d be on board with AI use (in game creation or otherwise) if a company were to say, “We’re not changing the budget we have for our human workforce, including for contractors, licensed art, and so on, other than increasing it as inflation and wages increase. We will be using ethical AI models to create more content than we otherwise would have been able to.” But I feel like in a corporate setting, its use is almost always going to result in them cutting jobs.


Are you okay with AAA studios using GenAI that was trained only on licensed works?


Okay, and? What nontechnical user cares enough to use it specifically when they could use Microsoft Office, Google Docs, Polaris Office, MobiOffice, WPS Office, Collabora, etc., instead?


But do nontechnical users care about the “missing” features? A lot of nontechnical users prefer simpler apps.
There is a version of Blender that was made for Android. It’s quite old, though. But if you’re competent enough with Blender that you’ve memorized all its keyboard shortcuts and workflows, you’re likely technical enough to get it working via Termux. But if not, Nomad Sculpt (on both iOS and Android), SpaceDraw (Android only), and several other apps can serve the same purposes.
Not sure why you listed video editing software and two different specific video editors, but Android and iOS both have Lumafusion. I’m sure there are other decent editors but I haven’t used them because Lumafusion is great. iPads do have DaVinci Resolve, though, for what that’s worth. If you care about using a FOSS video editor then you should care enough to install it via Termux. But let’s be real, most nontechnical users are probably happy using CapCut.
DJ software - Cross DJ is free. There are other alternatives. And there are web based DJ software apps like YouDJ.


OnlyOffice is available on Android already.
“any linux app” - I don’t think any nontechnical users want GParted on their Android phones, and it wouldn’t work anyway.
Android has its own games, same as iOS. Nontechnical users are way more likely to want Windows games than Linux games anyway.
Wine used to be developed natively for Android but they stopped a few years back. You can still download it at winehq though. I think Box64 with wine is a decent option?
Overall the thing I’m confused about is why you think Google or any major Android phone manufacturer have a motivation to make native Linux apps more accessible. Google certainly doesn’t want to make it easier for you to use the better versions of their competitors’ apps. Google is moving further away from Linux, not closer. Providing a usable, good enough desktop experience that’s still Android underneath makes far more sense for them.
Fortunately, like I said earlier, there are workarounds to get access to those Linux apps.
The thing that is more likely to change is for the creators of Android apps to build apps that function better when used in a phone-as-desktop format. And even if they don’t, there are enough competent web apps out there that just being able to use your browser full screen on a monitor solves 90% of people’s actual use cases - and probably over 95% when you include the other apps that have decent desktop experiences that can be run alongside them.
The Steam Deck approach is much closer to what you seem to want. The Steam Deck is an actually competent Linux machine that has a Valve-supported compatibility layer in Proton for running non-Linux games. It plugs into a USB-C hub connected to a monitor, mouse, and keyboard just fine, can install any Linux app, etc… It’s completely usable handheld as well. But it isn’t a phone, and even though it’s quite portable, it’s not “stick into your pocket” portable.
I don’t expect a major manufacturer to make a Linux phone any time soon, and I don’t think the Linux phones that are out already have - or will have in the next 5 years - a smooth enough experience to convince any nontechnical user to switch.


There are mobile versions for all of those?


What are the gaps in functionality for nontechnical people? And “apps that exist on Linux but not Android” doesn’t count, because such people are unlikely to have ever even used a Linux desktop in the first place. The improvement that matters won’t be Linux apps; it’ll be Android apps that are more usable in desktop mode.
That said, what are the issues with the apps that are currently available?
If a user installed Chrome, an office suite (whether that be Google Docs, Sheets, and Slides, the Microsoft equivalent, or something else), an email client, and other commonly available apps, what tasks would they be unable to complete, if any?
Are these, or other commonly used apps, substantially less usable than on desktop? If so, how so?


Can’t you just use GNURoot Debian and XServer SDL to get a Linux desktop env on any Android phone?
There’s an xda-developers guide on this and the two apps are still in the Google Play Store, so I assume it’s still feasible.
I’m not sure how well it plays with DeX and other similar solutions, though.
That’s assuming the apps aren’t capable enough to handle being used on a desktop on their own, of course. What sorts of gaps did you see, and in which sorts of apps?
The fisherman is being fished by someone above the panel and the bait is the iPhone.