For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse, and a bunch of oom killers start killing off various processes. In any case, I cannot successfully run the installer or Firefox by themselves.
https://bugzilla.redhat.com/show_bug.cgi?id=1559176
On Thu, Mar 22, 2018 at 4:30 AM, Chris Murphy lists@colorremedies.com wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse, and a bunch of oom killers start killing off various processes. In any case, I cannot successfully run the installer or Firefox by themselves.
For reference there's a GNOME performance hack fest [1] happening in May to review the stack to improve performance on low resource devices such as the Raspberry Pi and similar cheap SBCs, this should help improve this across all similar device classes/specs.
[1] https://wiki.gnome.org/action/show/Hackfests/Performance2018
On Thu, 2018-03-22 at 04:40 +0000, Peter Robinson wrote:
... to improve performance on low resource devices ...
Hi, I only recently, like three weeks ago, had been asked to update Fedora on a very old Dell notebook. I installed there Fedora 20 years ago and when I've been asked to get there a newer Firefox, because that ancient doesn't work for sites like Facebook (do not ask, it's not my machine), then I decided to update to the recent Fedora 27. Then the user gets all the security fixes and so on, right? Everyone wins. After successful 'distro-sync', which was quite impressive on its own, the gdm didn't boot quicker than in like 5 minutes, when I was lucky. Booting to GNOME took even longer. The rescue mode booted quicker, but that's not the production. There were caught some ABRT reports about kernel crash, but it was just about: "hey, kernel crashed, but you've bad luck, because from this crash kernel developers won't get anything, thus I'll not let you report it". What are such ABRT reports good for? The ancient Fedora 20 didn't have this issue, it just worked. Long story short, I spent half night and quite few hours the next day trying fresh install of Fedora 27, 26, 25, but none worked that well as Fedora 20, thus I resulted in getting the ancient Fedora 20 from the archives, install it and bring Firefox straight from Mozilla. The task accomplished, but...
Maybe there's some regression there, kernel should not crash, but I do not have that machine anymore, I've it only borrowed to get there new- enough browser and then I returned it back. It's not helpful for the community, I know, but it's the user experience I've got with one low resource device (I'm pretty sure it had the CPU and HDD higher than the minimum, the RAM at least 1GB too, but I'm not sure right now; did Fedora 20 have the same requirements? It works there reasonably well...).
Maybe you can compare such ancient Fedora, like the Fedora 20, with the current Fedora on those low resource devices to see (and feel) the difference. With classic "rotated disks", not SSD, that might be important too. Bye, Milan
On Thu, Mar 22, 2018 at 8:44 AM, Milan Crha mcrha@redhat.com wrote:
On Thu, 2018-03-22 at 04:40 +0000, Peter Robinson wrote:
... to improve performance on low resource devices ...
Hi,
I only recently, like three weeks ago, had been asked to update Fedora on a very old Dell notebook. I installed there Fedora 20 years ago and when I've been asked to get there a newer Firefox, because that ancient doesn't work for sites like Facebook (do not ask, it's not my machine), then I decided to update to the recent Fedora 27. Then the user gets all the security fixes and so on, right? Everyone wins. After successful 'distro-sync', which was quite impressive on its own, the gdm didn't boot quicker than in like 5 minutes, when I was lucky. Booting to GNOME took even longer. The rescue mode booted quicker, but that's not the production. There were caught some ABRT reports about kernel crash, but it was just about: "hey, kernel crashed, but you've bad luck, because from this crash kernel developers won't get anything, thus I'll not let you report it". What are such ABRT reports good for?
Well, what do you suggest? Not to inform user at all? ABRT tries to recognize if the report has some useful information, that would help to resolve this issue. Users are happy to click on button and create bugzillas like "I don't know what happened and have no idea how to reproduce". If the data itself do not have any info, then it is just noise for maintainers.
We, of course, could not inform users at all about some problems, but on the other hand - you would boot and computer would act strangely and you would have no idea if it is kernel/gnome/systemd/... problem. Now you knew that there is something wrong with kernel and you have a good starting point for you later investigation. (Or still, if you are experienced enough you can take closer look into the problem details)
The ancient Fedora 20 didn't have this issue, it just worked. Long story short, I spent half night and quite few hours the next day trying fresh install of Fedora 27, 26, 25, but none worked that well as Fedora 20, thus I resulted in getting the ancient Fedora 20 from the archives, install it and bring Firefox straight from Mozilla. The task accomplished, but...
Maybe there's some regression there, kernel should not crash, but I do not have that machine anymore, I've it only borrowed to get there new- enough browser and then I returned it back. It's not helpful for the community, I know, but it's the user experience I've got with one low resource device (I'm pretty sure it had the CPU and HDD higher than the minimum, the RAM at least 1GB too, but I'm not sure right now; did Fedora 20 have the same requirements? It works there reasonably well...).
Maybe you can compare such ancient Fedora, like the Fedora 20, with the current Fedora on those low resource devices to see (and feel) the difference. With classic "rotated disks", not SSD, that might be important too. Bye, Milan _______________________________________________ desktop mailing list -- desktop@lists.fedoraproject.org To unsubscribe send an email to desktop-leave@lists.fedoraproject.org
On Thu, 2018-03-22 at 09:52 +0100, Matej Marusak wrote:
Now you knew that there is something wrong with kernel and you have a good starting point for you later investigation.
Hi, yes, I agree, informing is good, but this one is not useless only for developers, it's useless also for the user. It's a starting point, but also the final point. I do not know how to debug kernel crashes and I'd bet the "Details" button was disabled (it's a long time, I can be wrong), not talking that getting information from the ABRT GUI (like the most important is the backtrace of the crash for me) is a problem, as it's not available there at all or it cannot be copied from the GUI (not everything in the GUI of ABRT can be copied out).
My options, as I see them:
- ABRT tells me kernel crashed during boot, but doesn't let me file
the bug
As a regular user: - file a bug in probably Red Hat bugzilla and hope kernel folks monitor it, thus they can help/guide me and so on - read about the crash the next boot, until...
As a developer: - oops, kernel crashed. I've no idea how to debug kernel crashes during boot, thus let's file a bug and continue as a regular user above.
So I end up with a bug and with a hope. What's the main difference between the bug filled by ABRT and the one filled by me? As long as the anonymous bug reports are disabled (it's the FAF for, right, bugs cannot be filled anonymously), then the bug I fill manually is even less useful than the one which ABRT would fill. Why? ABRT may include basic information about the issue, like what crashed, when/where it crashed, what versions of packages had been involved and all that information which the developer would ask for anyway, but which I'd not always be able to give him/her easily (see above, cannot copy everything from the ABRT GUI). Thus ABRT-filled bug would be more accurate and contain much more information than what the user would do. Frankly, my initial bug text would probably look like:
ABRT tells me about a crash of kernel after boot, but ABRT claims the report would be useless for kernel developers, thus it doesn't let me to report it. Could you guide me what to do now to investigate the crash, please? It's there after each restart of the machine.
I even do not know what heuristic is done to make ABRT think that the human kernel developer would consider the report useless.
(Or still, if you are experienced enough you can take closer look into the problem details)
This is kernel/boot related. Most users are not experienced enough, I'm pretty sure of it.
Bye, Milan
On Thu, Mar 22, 2018 at 10:24 AM, Milan Crha mcrha@redhat.com wrote:
On Thu, 2018-03-22 at 09:52 +0100, Matej Marusak wrote:
Now you knew that there is something wrong with kernel and you have a good starting point for you later investigation.
Hi,
yes, I agree, informing is good, but this one is not useless only for developers, it's useless also for the user. It's a starting point, but also the final point. I do not know how to debug kernel crashes and I'd bet the "Details" button was disabled (it's a long time, I can be wrong), not talking that getting information from the ABRT GUI (like the most important is the backtrace of the crash for me) is a problem, as it's not available there at all or it cannot be copied from the GUI (not everything in the GUI of ABRT can be copied out).
Details button being disabled is weird. If it was so, it may be that there were no details? If it is really that case that there were no details at all, then I agree with you - we should not bother users with it. I cannot recall ever seeing that, I also connected with few friend who use Fedora and are nagging that kernel is always crashing and all of the reports had details enabled.
That not everything can be copied out of GUI - well I don't like the GUI as well and I would love to rewrite it from scratch. If I want to have GUI I use Cockpit [1] (it does not yet support reporting).
[1] https://abrt.github.io/abrt/cockpit/2017/06/29/ABRT-in-cockpit/
My options, as I see them:
- ABRT tells me kernel crashed during boot, but doesn't let me file
the bug
As a regular user: - file a bug in probably Red Hat bugzilla and hope kernel folks monitor it, thus they can help/guide me and so on - read about the crash the next boot, until...
I see it differently - regular user sees problem, if everything works then usually is not bothered at all. If is more active, then hits report button.
As a developer: - oops, kernel crashed. I've no idea how to debug kernel crashes during boot, thus let's file a bug and continue as a regular user above.
Yes developer would probably hit the report button or even create bugzilla.
So I end up with a bug and with a hope. What's the main difference between the bug filled by ABRT and the one filled by me? As long as the anonymous bug reports are disabled (it's the FAF for, right, bugs cannot be filled anonymously), then the bug I fill manually is even less useful than the one which ABRT would fill. Why? ABRT may include basic information about the issue, like what crashed, when/where it crashed, what versions of packages had been involved and all that information which the developer would ask for anyway, but which I'd not always be able to give him/her easily (see above, cannot copy everything from the ABRT GUI). Thus ABRT-filled bug would be more accurate and contain much more information than what the user would do.
I must repeat myself - point of disabling reporting is not to storm developers with version of package and command (for which we have FAF, where you can see how much reports are in which version of package, you can see there also command etc...). Bugzilla should contain at least reproducer/coredump or something that can be used for tracing what happened.
However I am interested in usability of ABRT (mainly FAF, but GUI as well) and we can talk about this more and come with what should be done. But let's continue out off this thread as it a little bit OT. Either contact me on mmarusak@redhat.com or on our mailing list abrt-devel-list@redhat.com
MM
Frankly, my initial bug text would probably look like:
ABRT tells me about a crash of kernel after boot, but ABRT claims the report would be useless for kernel developers, thus it doesn't let me to report it. Could you guide me what to do now to investigate the crash, please? It's there after each restart of the machine.
I even do not know what heuristic is done to make ABRT think that the human kernel developer would consider the report useless.
(Or still, if you are experienced enough you can take closer look into the problem details)
This is kernel/boot related. Most users are not experienced enough, I'm pretty sure of it.
Bye, Milan
desktop mailing list -- desktop@lists.fedoraproject.org To unsubscribe send an email to desktop-leave@lists.fedoraproject.org
2018-03-22 1:44 GMT-06:00 Milan Crha mcrha@redhat.com:
On Thu, 2018-03-22 at 04:40 +0000, Peter Robinson wrote:
... to improve performance on low resource devices ...
Hi,
I only recently, like three weeks ago, had been asked to update Fedora on a very old Dell notebook. I installed there Fedora 20 years ago and when I've been asked to get there a newer Firefox, because that ancient doesn't work for sites like Facebook (do not ask, it's not my machine), then I decided to update to the recent Fedora 27. Then the user gets all the security fixes and so on, right? Everyone wins. After successful 'distro-sync', which was quite impressive on its own, the gdm didn't boot quicker than in like 5 minutes, when I was lucky. Booting to GNOME took even longer. The rescue mode booted quicker, but that's not the production. There were caught some ABRT reports about kernel crash, but it was just about: "hey, kernel crashed, but you've bad luck, because from this crash kernel developers won't get anything, thus I'll not let you report it". What are such ABRT reports good for? The ancient Fedora 20 didn't have this issue, it just worked. Long story short, I spent half night and quite few hours the next day trying fresh install of Fedora 27, 26, 25, but none worked that well as Fedora 20, thus I resulted in getting the ancient Fedora 20 from the archives, install it and bring Firefox straight from Mozilla. The task accomplished, but...
I a computer without a SSD Plasma and Gnome Shell do not work well, so in the current state os those desktop enviroments the page also should recomend to use a SSD disck, is not use a alternative desktop like Cinnamon or a more clasical option like Mate.
On Wed, Mar 21, 2018 at 10:30 PM, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse, and a bunch of oom killers start killing off various processes. In any case, I cannot successfully run the installer or Firefox by themselves.
Comparing Fedora 27 ISO to Fedora-Workstation-Live-x86_64-28-20180320.n.0.iso, the biggest difference is Fedora 28 packagekitd is soaking the CPU nearly 100% upon initial login and uses upwards of 20% of the available memory before settling down to 10%.. This is without launching anything. Whereas on Fedora 27, the memory footprint for packagekitd is 0.2%.
Fedora 27 ISO I can launch and use Firefox with a VM of 1.5G. I can't with Fedora 28, and the VM reports dozens of: Mar 22 00:47:49 localhost-live org.gnome.Shell.desktop[1707]: Window manager warning: HW cursor for format 875713089 not supported
And then just becomes unresponsive.
Anyway, tentatively it looks like 1G was a bogus minimum for Fedora 27, where 1.5G would have worked. And for Fedora 28 it needs to be 2G.
On Wed, Mar 21, 2018 at 10:30:09PM -0600, Chris Murphy wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse
systemd-coredump should probably short-circuit the coredump in oom conditions. I'm not sure what the exact conditions should be, maybe if the amount of free RAM is lower then the core size?
Zbyszek
----- Original Message -----
On Wed, Mar 21, 2018 at 10:30:09PM -0600, Chris Murphy wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse
systemd-coredump should probably short-circuit the coredump in oom conditions. I'm not sure what the exact conditions should be, maybe if the amount of free RAM is lower then the core size?
Why does the kernel even make that information available? I regularly have my machine made unusable when the kernel decides to kill the largest memory users (usually the mail client, or a browser tab), and it OOM kills the largest memory users, meaning the ones that will take the longest to dump to disk.
We already have records of the application being killed through OOM in the journal, we should be spared even trying to pass the core to user-space.
On Thu, Mar 22, 2018 at 07:07:12AM -0400, Bastien Nocera wrote:
----- Original Message -----
On Wed, Mar 21, 2018 at 10:30:09PM -0600, Chris Murphy wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse
systemd-coredump should probably short-circuit the coredump in oom conditions. I'm not sure what the exact conditions should be, maybe if the amount of free RAM is lower then the core size?
Why does the kernel even make that information available? I regularly have my machine made unusable when the kernel decides to kill the largest memory users (usually the mail client, or a browser tab), and it OOM kills the largest memory users, meaning the ones that will take the longest to dump to disk.
We already have records of the application being killed through OOM in the journal, we should be spared even trying to pass the core to user-space.
That depends. Sometimes you want the core on oom to diagnose a memory leak. There are two different cases: - a single program went awry and used all memory and was killed. After it is gone, the machine will be usable again. It's potentially useful to dump the core. - the machine is generally starved for resources, and even after the biggest consumer of memory has been killed, there still is not enough. Dump of the core puts further strain on the system and should be avoided.
The question is how to distinguish those two cases.
Zbyszek
----- Original Message -----
That depends. Sometimes you want the core on oom to diagnose a memory leak. There are two different cases:
- a single program went awry and used all memory and was killed. After it is gone, the machine will be usable again. It's potentially useful to dump the core.
- the machine is generally starved for resources, and even after the biggest consumer of memory has been killed, there still is not enough. Dump of the core puts further strain on the system and should be avoided.
The question is how to distinguish those two cases.
I can't think of a single time of when I'd want to wait for the core dumping to finish (which can make the machine unusable for tens of minutes) to debug a memory usage problem.
I'd say making this a kernel tunable that's disabled by default would cover the vast majority of uses, development machines included.
On Thu, Mar 22, 2018 at 09:45:39AM -0400, Bastien Nocera wrote:
I can't think of a single time of when I'd want to wait for the core dumping to finish (which can make the machine unusable for tens of minutes) to debug a memory usage problem.
I can think of many times — but none of them are individual desktop use cases. This could be very valuable in server, HPC, or even enterprise desktop contexts.
On Mon, Mar 26, 2018 at 08:15:36AM -0400, Matthew Miller wrote:
I can't think of a single time of when I'd want to wait for the core dumping to finish (which can make the machine unusable for tens of minutes) to debug a memory usage problem.
I can think of many times — but none of them are individual desktop use cases. This could be very valuable in server, HPC, or even enterprise desktop contexts.
(So, given the Fedora Workstation target, I agree with you. We probably don't want this on by default in this edition.)
----- Original Message -----
On Thu, Mar 22, 2018 at 09:45:39AM -0400, Bastien Nocera wrote:
I can't think of a single time of when I'd want to wait for the core dumping to finish (which can make the machine unusable for tens of minutes) to debug a memory usage problem.
I can think of many times — but none of them are individual desktop use cases. This could be very valuable in server, HPC, or even enterprise desktop contexts.
Dumping the core of the application that might not even be the one responsible for eating all the RAM? Please expand.
Enterprise desktop users are the same as desktop users, except that they can't afford to take a 30-minute break when their machine is thrashing. If you know any different, I'm all for an expanded explanation here.
On Wed, Mar 28, 2018 at 04:53:20AM -0400, Bastien Nocera wrote:
Dumping the core of the application that might not even be the one responsible for eating all the RAM? Please expand.
Enterprise desktop users are the same as desktop users, except that they can't afford to take a 30-minute break when their machine is thrashing. If you know any different, I'm all for an expanded explanation here.
If I'm having a persistent but difficult to trace problem across my set of X hundred or thousand managed workstations, I want all the information I can get. The downtime would be no different than if a hard drive failed or something — if it's half an hour, sorry. If it's more than that, that's why we have a pool of loaner machines.
----- Original Message -----
On Wed, Mar 28, 2018 at 04:53:20AM -0400, Bastien Nocera wrote:
Dumping the core of the application that might not even be the one responsible for eating all the RAM? Please expand.
Enterprise desktop users are the same as desktop users, except that they can't afford to take a 30-minute break when their machine is thrashing. If you know any different, I'm all for an expanded explanation here.
If I'm having a persistent but difficult to trace problem across my set of X hundred or thousand managed workstations, I want all the information I can get. The downtime would be no different than if a hard drive failed or something — if it's half an hour, sorry. If it's more than that, that's why we have a pool of loaner machines.
How is that a different behaviour from setting, say, a system through which you can dump kernel cores, or any other root causing done by an admin?
I said:
I'd say making this a kernel tunable that's disabled by default would cover the vast majority of uses, development machines included.
And that's still what we would want. Especially given that the core that's dumped in the end might not even be that of the culprit.
On Wed, Mar 28, 2018 at 10:50:41AM -0400, Bastien Nocera wrote:
I'd say making this a kernel tunable that's disabled by default would cover the vast majority of uses, development machines included.
And that's still what we would want. Especially given that the core that's dumped in the end might not even be that of the culprit.
I wasn't trying to argue. I agree that this is the right choice for Workstation.
On Do, 22.03.18 08:46, Zbigniew Jędrzejewski-Szmek (zbyszek@in.waw.pl) wrote:
On Wed, Mar 21, 2018 at 10:30:09PM -0600, Chris Murphy wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
https://getfedora.org/en/workstation/download/
When I set a VM to 1G it's always unusable. Sometimes it's PackageKit that crashes first, which causes systemd-coredump to spin up and run out of memory next, and then oom killers start killing off various processes. Sometimes the oom killing just starts happening without a prior crash. In any case I never get a desktop.
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?), it kicks off systemd-coredump which makes the memory problem worse
systemd-coredump should probably short-circuit the coredump in oom conditions. I'm not sure what the exact conditions should be, maybe if the amount of free RAM is lower then the core size?
Hmm, to be frank, I wasn't aware the kernel would trigger coredump handling in case of OOM. This feels a bit like making things better and then making things worse again ;-)
I wonder if there's any way to detect the reason the kernel calls the coredump handlers, i.e. that it can tell us whether OOM triggered it or anything else.
Lennart
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the problem is going to be libsolv rather than much we can fix in "Fedora".
Richard.
On Thu, Mar 22, 2018 at 3:15 AM, Richard Hughes hughsient@gmail.com wrote:
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the problem is going to be libsolv rather than much we can fix in "Fedora".
Never used massif, don't know how to use it, but I'll do whatever I'm told on any machine I have, real or VM.
What I'm noticing is if the VM is set to 1G RAM I pretty much always see packagekitd crash. And it's too memory limited of an environment for a live to then get any data out of it, even serial console with 'virsh console' hangs up for tens of minutes at a time. If I go to 1.5G RAM the packagekitd crash I'd say is rare, but the sample size so far is maybe 10 boots. I don't know if what you need is a crashed packagekitd to find out why it's crashing, or something else.
On Thu, Mar 22, 2018 at 3:15 AM, Richard Hughes hughsient@gmail.com wrote:
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the problem is going to be libsolv rather than much we can fix in "Fedora".
OK I got it to crash in a way that I was able to collect some minimal info...
https://bugzilla.redhat.com/show_bug.cgi?id=1559684
There's a link to the coredump there. For whatever reason the retrace server will not accept it with abrt. https://github.com/abrt/retrace-server/issues/180
On 03/23/2018 04:06 AM, Chris Murphy wrote:
On Thu, Mar 22, 2018 at 3:15 AM, Richard Hughes hughsient@gmail.com wrote:
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the problem is going to be libsolv rather than much we can fix in "Fedora".
OK I got it to crash in a way that I was able to collect some minimal info...
https://bugzilla.redhat.com/show_bug.cgi?id=1559684
There's a link to the coredump there. For whatever reason the retrace server will not accept it with abrt. https://github.com/abrt/retrace-server/issues/180
Thanks Chris. The coredump seems to be just various things starting to go wrong in packagekitd when we start running out of resources and I don't think there's much we can do there, HOWEVER:
The main issue seems to be that we're starting gnome-software on the live media, which in turn is starting packagekitd and asking it to download metadata. This is 100 MB download which eats up precious RAM as the file system overlay on the live media is backed by memory.
This all adds up with gnome-software using RAM, packagekitd using RAM, and downloaded metadata using RAM.
Working on fixing it now to avoid all that on live media.
On Fri, 2018-03-23 at 11:37 +0100, Kalev Lember wrote:
On 03/23/2018 04:06 AM, Chris Murphy wrote:
On Thu, Mar 22, 2018 at 3:15 AM, Richard Hughes hughsient@gmail.com wrote:
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the problem is going to be libsolv rather than much we can fix in "Fedora".
OK I got it to crash in a way that I was able to collect some minimal info...
https://bugzilla.redhat.com/show_bug.cgi?id=1559684
There's a link to the coredump there. For whatever reason the retrace server will not accept it with abrt. https://github.com/abrt/retrace-server/issues/180
Thanks Chris. The coredump seems to be just various things starting to go wrong in packagekitd when we start running out of resources and I don't think there's much we can do there, HOWEVER:
The main issue seems to be that we're starting gnome-software on the live media, which in turn is starting packagekitd and asking it to download metadata. This is 100 MB download which eats up precious RAM as the file system overlay on the live media is backed by memory.
This all adds up with gnome-software using RAM, packagekitd using RAM, and downloaded metadata using RAM.
Working on fixing it now to avoid all that on live media.
Can you or Chris propose a bug for this as a Beta freeze exception and give it an accurate description (like 'GNOME Software runs unnecessarily on live images' or so?) It'd be good to get a fix for this into Beta I guess.
I do still suspect anaconda is using more memory than before as well, but will have to look into that, and it may not affect liveinst as the live installer doesn't use the dnf payload.
On Sat, 24 Mar 2018, 01:23 Adam Williamson, adamwill@fedoraproject.org wrote:
On Fri, 2018-03-23 at 11:37 +0100, Kalev Lember wrote:
On 03/23/2018 04:06 AM, Chris Murphy wrote:
On Thu, Mar 22, 2018 at 3:15 AM, Richard Hughes hughsient@gmail.com
wrote:
On 22 March 2018 at 04:30, Chris Murphy lists@colorremedies.com
wrote:
With 1.5G RAM I usually get a desktop but then if anything crashes, and it seems like PackageKit is crashing often when under memory pressure (?)
PackageKit-the-supervisor uses very little memory indeed by design, but PackageKit does also load libdnf which loads libsolv etc which is probably where the problem lays. Do you have a massif run handy on a small machine? I can do one myself if not, although I fear the
problem
is going to be libsolv rather than much we can fix in "Fedora".
OK I got it to crash in a way that I was able to collect some minimal
info...
https://bugzilla.redhat.com/show_bug.cgi?id=1559684
There's a link to the coredump there. For whatever reason the retrace server will not accept it with abrt. https://github.com/abrt/retrace-server/issues/180
Thanks Chris. The coredump seems to be just various things starting to go wrong in packagekitd when we start running out of resources and I don't think there's much we can do there, HOWEVER:
The main issue seems to be that we're starting gnome-software on the live media, which in turn is starting packagekitd and asking it to download metadata. This is 100 MB download which eats up precious RAM as the file system overlay on the live media is backed by memory.
This all adds up with gnome-software using RAM, packagekitd using RAM, and downloaded metadata using RAM.
Working on fixing it now to avoid all that on live media.
Can you or Chris propose a bug for this as a Beta freeze exception and give it an accurate description (like 'GNOME Software runs unnecessarily on live images' or so?) It'd be good to get a fix for this into Beta I guess.
I do still suspect anaconda is using more memory than before as well, but will have to look into that, and it may not affect liveinst as the live installer doesn't use the dnf payload.
I agree here, I suspect, although I've not investigated, that the initrd contents has bloated due some random increase in deps causing memory used by the ram disk to increase.
On 03/23/2018 06:23 PM, Adam Williamson wrote:
Can you or Chris propose a bug for this as a Beta freeze exception and give it an accurate description (like 'GNOME Software runs unnecessarily on live images' or so?) It'd be good to get a fix for this into Beta I guess.
Done: https://bugzilla.redhat.com/show_bug.cgi?id=1560504
On Wed, 2018-03-21 at 22:30 -0600, Chris Murphy wrote:
For Fedora 27 it says this:
Fedora Workstation to a laptop or desktop computer that has at least 1 GHz processor, 1 GB RAM, and 10 GB space available
There's no 'review process' for this text, it just gets copied from release to release. The fact that it's there is not really a guarantee that the release *actually* worked in those specs.
I remember a few cycles back it still said 256MB or 512MB or something and I noticed and jumped through some hoops to get it changed to 1GB.
So, as you discovered, it'd be better to actually *test* F27 and F28 in the same ways and see how they compare.
openQA actually gathers some data on used memory, but unfortunately the numbers for the F27 release are lost (I didn't have things set up properly to protect release candidate results from garbage collection at the time). Logs of installer memory use in the 'compose check report' emails from 20171105 mention peak usage on the Workstation network installer as 582MiB , whereas the number from yesterday's compose appears to be about 850MiB. The peak usage is early in the run, which I think indicates it's occurring during repository retrieval.
Combined with your numbers on PackageKit memory usage, that sorta indicates to me that something down in the stack that both dnf (used in anaconda) and PackageKit use is sucking up a lot more memory than it used to. That'd be worth looking into in more detail, if you can (I'm a lil' busy right now).
desktop@lists.fedoraproject.org