this post was submitted on 10 Dec 2023
244 points (87.0% liked)

AssholeDesign

7564 readers
3 users here now

This is a community for designs specifically crafted to make the experience worse for the user. This can be due to greed, apathy, laziness or just downright scumbaggery.

founded 1 year ago
MODERATORS
 

there's just this vague recommendation to "turn it off in settings." no link, no toggle, nothing.

had to hunt for the relevant setting in the app(even their settings menu is down under in the list when you tap on your avatar.


image transcription:

a screenshot of Google play store which has a popup that can only be disabled by pressing OK button at the bottom.
the popup contains the following text:

Google is optimising app installs with your help
Google Play makes apps faster to install, open and run based on what people are using most. The first time that you open an app after installing, Google notes which parts of the app you use. When enough people do this, your apps will install faster. App install optimisation is automatically turned on, but you can turn it off in settings. Learn more

all 37 comments
sorted by: hot top controversial new old
[–] henfredemars@infosec.pub 239 points 11 months ago (5 children)

Nobody here seems to understand the real technical purpose of this feature, so let me try to explain what's really happening.

Android applications and much of the system are written in Java. In the beginning, this decision was made because it provides portability at least in theory and the language was familiar to developers, but this decision wasn't without its disadvantages. This extra layer of the Java virtual machine led to laggy user experiences and increased memory use in the very first versions of Android.

The first attempt to fix this problem was the Dalvik runtime. This first solution incorporates a tracing style just in time compiler that translates Java into native code as it runs. An Android Kitkat, ART or Android runtime was introduced as a developer option and later became the default which offers full compilation to native code at install time. This led to a really long system update process and some complex apps took forever to install, so in Nougat Android moved to a hybrid design that only pre-compiles the most important parts of the application. This saves install time and system update time while still getting good performance by pre-compiling to native, and we fall back to JIT if unexpected code becomes a performance bottleneck when the user is actually using the app.

But which parts do we pre-compile? How do you know what code actually matters to the user experience?

This dialogue in Play is about sharing runtime information about which code was actually executed. Google aggregates this information so that when users install an app they also receive a file describing which parts are most commonly used and actually matter to performance so the system can focus on pre-compiling only this specific subset. Namely, this data is which classes and methods are expected to be hot, with special attention paid to the ones required to initialize the application.

From a technical perspective, I don't really see how this places private user information at risk, and I agree that it needs to be the default for this design to work because most users aren't savvy enough to understand what it actually does. Google tried to simplify it is much as possible in this dialogue, but it's led to a lot of confusion especially for more technical users.

[–] SeedyOne@lemm.ee 39 points 11 months ago

The detailed breakdown is appreciated.

[–] r00ty@kbin.life 19 points 11 months ago (1 children)

Silly question, I'm a software dev but not as mobile dev so could be barking up the wrong tree here. Could Google play not cache precompiled versions for the most popular architectures and just download them precompiled? Sure some rarer configurations would still need to do the slower local process but the more popular configurations representing the vast majority of systems would be covered.

[–] henfredemars@infosec.pub 28 points 11 months ago* (last edited 11 months ago) (4 children)

It's not impossible, but there are limitations to that alternative approach. For example, ARM processors are really heterogeneous with respect to exactly which extensions are supported. If you compile on the server, you might not be able to take advantage of features that only exist on a few phones like the latest flagships. It also requires the device to trust that the pre-compiled code is actually consistent and won't cause bad memory access for example that can lead to a crash. If the device compiles the code, it knows there's a true one-to-one correspondence with the actual Java byte code, closing an opportunity for inconsistency. The device knows the size of the CPU caches and might make different instruction scheduling decisions in the actual binary. Generally, it's easier to be more efficient when you have all the information about the target processor available on device. But, you could have done it a different way that's true. Google decided not to do that.

We should also think about what's to be gained by moving it all to the server? On the server, you still don't know the actual user patterns. It's not space efficient to compile the entire application even if the compile time is free on the server. You would have to download unnecessary native code that may only be used by an extremely small fraction of users, wasting storage space on user devices. What if I don't use that whole area in the application? What if most users don't use it? Why should we have to download an optimized version of it onto every device? Perhaps, the best optimization is to not ship any native code for the feature nobody uses.

If you want the best balance of high performance on the user device while not wasting extra space for binary code that isn't frequently used, I think this design makes more sense. You're never getting away from having a full compiler on every device because Google can't assume that you aren't using other app stores, and it's perfectly valid for Android apps to generate code dynamically at runtime that can never be compiled in advance.

[–] Unpigged@lemmy.dbzer0.com 8 points 11 months ago

I want to appreciate the effort you put in this response. Respect, dear sir.

[–] lazyvar@programming.dev 6 points 11 months ago (1 children)

I feel you’re brushing over the privacy implications regarding how apps are used.

Sure, you could say: “Oh, but it’s inefficient to compile the entire application, and what if there are features that barely anyone uses.”

But you can also say: “Compiling the entire application ensures we don’t need to collect usage data and it ensures everyone gets the best experience, even the people that use features that are otherwise hardly used.”

Now, of course, to go with the second option, you need to care about user privacy and not gain any benefits from usage data beyond the benefits for compiling it.

[–] henfredemars@infosec.pub 6 points 11 months ago

You're right. That is a perfectly valid solution with different trade offs. Perhaps in a perfect world users could select which they prefer.

[–] r00ty@kbin.life 2 points 11 months ago (1 children)

Yes, it's why I said most popular configurations as well as architecture, not just CPU architecture. So that's specific CPU/SoC components (GPU etc). Compiling for the most popular configurations (thinking Google Pixel, Samsung Sxx, whatever Motorola's flagship is. Probably also the mid-range units that sell well too) you would cover a large %age with maybe 10-15 configurations, which isn't THAT hard to do if you're google.

[–] henfredemars@infosec.pub 5 points 11 months ago

This is the kind of approach that Apple takes by building multiple versions of the application for different devices. It's a perfectly serviceable solution that also works.

I don't have a great answer for why they don't do it this way.

[–] space@lemmy.dbzer0.com 1 points 11 months ago (1 children)

While not all ARMs are the same, I'm pretty sure there are certain configurations that are more common. Why not just provide binaries for those?

[–] henfredemars@infosec.pub 1 points 11 months ago

You can, but that could be to the detriment of less common configurations. That's a valid solution that strikes a different balance.

You may also wish to consider that Android apps can generate and execute code at runtime. It's perfectly valid to download or generate more Java bytecode after the app was installed. You would still need a compiler on the device to handle apps that do this efficiently.

I also left out some more specific information to prevent my reply from being even longer, but the binary code your device generates actually adapts to your specific usage as well as the general usage. The profile you get from Play is just a starting point. Shipped binary code isn't generally very flexible. If you like to use an app in a way that makes us runtime compile code that we thought wasn't important before, your device actually just adds this information to the profile and recompiles when the device isn't being used. That means you get a personal level of performance optimization depending on your personal usage.

It's not wrong to ship a slightly less optimal, fully compiled binary. I just don't think that would work as well. It would be a lot simpler though.

[–] HelloHotel@lemmy.world 4 points 11 months ago* (last edited 11 months ago) (1 children)

My only question now is, What parts are in AOSP and what parts are Googled Android? Is reading and installing with these optomization files/API sys-app only? Are they using accessability to track thd open activity, or via the optomization API provided by android.

I cant really tell you if their acting in bad faith without this info.

[–] henfredemars@infosec.pub 13 points 11 months ago* (last edited 11 months ago) (1 children)

I've done a dive of the source code and I can provide a partial answer.

The execution logging functionality is publicly available source code right here.. The code reveals how Android decides what and when to save hot methods.

Dex2oat uses the profile data. You can see references here to the profile directory.

Now, where is the code that ships and downloads these profiles? I can't find it. I suspect it's part of Play services. Thus, AOSP provides all the machinery, but Play probably handles the actual profile data delivery.

[–] HelloHotel@lemmy.world 1 points 11 months ago

Thats what I feared. Its either ptrace, root file access or a non public android API

[–] lemmesay@discuss.tchncs.de -5 points 11 months ago (1 children)

I knew it had something to do with art but didn't know you could pre-compile only certain segments. I thought the best you could do was to have a bundle for certain architecture (split apks). thanks for the explanation!

now I have another question:
can't Google just get that data by running the app itself (which it does anyway) on the servers? or do they trust real hardware more than emulators?

[–] henfredemars@infosec.pub 16 points 11 months ago

Google cannot do this because the code that matters to performance depends on all the inputs which includes the user's actual behavior. For example, I could write an app that contains a big feature that users overwhelmingly care nothing about, wasting time in compiling it and resources in storing the translated code for this unused feature.

Building on the server means giving up access to user behavior, which is important to what code really runs. It's tough to strike a good balance between not storing code that won't be used, install times, and runtime performance.

[–] Showroom7561@lemmy.ca 51 points 11 months ago (1 children)

"Apps will install faster."

Look. I'm using a 4 year old phone, and apps seem to install in seconds. Downloading them takes longer than the install, and even that's a few seconds...

What problem is Google really trying to solve here?

[–] Denvil@lemmy.one 27 points 11 months ago

They're trying to solve their money problems, we all know such a small company like Google must be struggling

[–] Tau@sopuli.xyz 28 points 11 months ago (1 children)

Seems like a cool idea but it would be good if it was opt-in and done by not-google

[–] Swarfega@lemm.ee 19 points 11 months ago* (last edited 11 months ago)

If they do that people won't turn it on. They rely on people to not read things.

I wish someone (EU) would step in and force companies to make data collection settings be opt in rather than opt out.

[–] manucode@infosec.pub 15 points 11 months ago (2 children)

I could neither find any such setting nor did I ever get such a notification. Could you tell me where you found the relevant setting?

[–] mexicancartel@lemmy.dbzer0.com 4 points 11 months ago (1 children)

I remember seeing this popup but cant find anything in settings

[–] trainden@lemmy.blahaj.zone 15 points 11 months ago (2 children)

It's in Settings > General > App install optimisation. The last item in the list.

[–] mexicancartel@lemmy.dbzer0.com 5 points 11 months ago* (last edited 11 months ago)

There is no such option for me. I guess its not aviliable

[–] manucode@infosec.pub 0 points 11 months ago

Found it and turned it off. Thanks

[–] henfredemars@infosec.pub 3 points 11 months ago* (last edited 11 months ago)

Not all devices are supported. Building a list of what code is actually run is going to be device dependent, so it only handles popular configurations.

Or, it could be that you're running an older version of the play store, or maybe there's some other reason that it's not using this feature that I don't know about.

[–] KazuyaDarklight@lemmy.world 8 points 11 months ago

So Google basically wants to add a feature to allow for running apps before they are fully installed? Same way you can play WoW after a minimum amount of the game has been downloaded, but before it's entirely done. Definitely feels like overkill for the majority of apps I use.

[–] thomasfrank@discuss.tchncs.de 2 points 10 months ago (1 children)

@pokedoku You're right, that popup is vague and unhelpful.

[–] monk@lemmy.unboiled.info 1 points 7 months ago

You forgot to degoogle.