- 🇫🇷France 5
I am also facing the issue on a taxonomy term page.
Before searching in core issue queue, I realized my MariaDB configuration was the default one and I upgraded it according to https://www.drupal.org/node/259580 → with a single effect: the "MySQL Server has gone away error" has been replaced by a PHP out of memory error (512 Mo allowed).
Patch from #15 has fixed the situation (thanks Defcon0).
- 🇫🇮Finland iSampo
Having this issue as well, MySQL goes away when it tries to cache all the site users (~12k) in
cache_entity
table, and then dblog does its job by trying to set a single log message for that insert with@message
that contains the 85k placeholders.The patch in #15 does get rid of the error but has issues, including all
cid
s being wrong in all cache tables due to not preserving keys.Here's a patch that does the following:
- Move the chunking intodoSetMultiple()
to have it done only once. Not really a fan of the looping the chunks and then the items but here it is anyways. No other logic changes inside the function. Doing it this way it will of course increase the number of inserts by quite much so not sure what kind of implications that would have.
- Preserve thecid
s
- UpdateMAX_ENTITIES_PER_CACHE_SET
toMAX_ITEMS_PER_CACHE_SET
- 🇫🇷France fgm Paris, France
Another take here 🐛 SqlContentEntityStorage::loadMultiple needs to limit the number of items stored at once Closed: duplicate .
- 🇫🇷France fgm Paris, France
FWIW, in the other issue I had to raise
max_allowed_packet = 1024M
to support the use case. Maybe we could look up the value of the parameter at some point, maybe during cache flushes/container rebuilds and adjust thatMAX_ITEMS_PER_CACHE_SET
constant accordingly: 100 seems low for most setups. - 🇬🇧United Kingdom catch
I don't think the 100 limit is much of a problem even on systems that can handle more, the ::setMultiple() is an optimisation against setting items one by one. Even with 1,000 items to set, that's only nine extra database queries, and each database query is likely to be faster given they're smaller.
- 🇫🇷France fgm Paris, France
Yeah, maybe just set it as a public or even protected property so that anyone with an actual need for higher values can tweak it from the outside (using setAccessible if it is protected).
Hi I have a website with about 30 Taxonomy, since 9.4.9 2 of my taxo crash, the two more big about 700 taxonomy, soo I applied this patch now works nice thx :)
But, I have a content Books, each content has about 100-200 pages, impossible to acces now and my View crach, 9.4.9 make Big limitations and Crash many many site! xD
Now Worpress do better for classification.... xD
What I can do? Thx
Bug correct the viewss for taxonomy, but Big content (about 100 pages) block, and in Back-end report the LOG is break, message:
"The website encountered an unexpected error. Please try again later."
All page with a bit data seem BUG, patch no correct this. :(
I find solution for the moment, I put on my View with BIG TAXO to show only 10 content and works, more 20 no works Error server....
- last update
over 1 year ago Custom Commands Failed - Status changed to Needs review
over 1 year ago 5:04pm 11 September 2023 - 🇬🇧United Kingdom catch
#21 seems OK to me, it would be worth adding explicit test coverage to the database cache backend that you can set at least 101 cache items.
- last update
over 1 year ago Custom Commands Failed - 🇸🇰Slovakia poker10
Added a new test case to the
DatabaseBackendTest::testSetGet()
to test exceeding the chunk size. No other changes made. - last update
over 1 year ago 30,148 pass - Status changed to Needs work
over 1 year ago 2:54pm 12 September 2023 - Status changed to Needs review
over 1 year ago 10:25pm 12 September 2023 - Status changed to RTBC
over 1 year ago 8:09pm 13 September 2023 - 🇺🇸United States smustgrave
Thanks @acbramley!
IS lines up to solution and test coverage seems to cover #31
- last update
over 1 year ago 30,150 pass - Status changed to Fixed
over 1 year ago 6:20am 14 September 2023 - 🇬🇧United Kingdom catch
Committed/pushed to 11.x and cherry-picked to 10.1.x, thanks!
Automatically closed - issue fixed for 2 weeks with no activity.
- Status changed to Fixed
about 1 year ago 4:51pm 1 October 2023 Patch 34 is still Oki for TAxo big, but problem is still here for show big content, all my COntent BIG is bloque with error:
Error 503 Backend fetch failed
Backend fetch failedGuru Meditation:
XID: 699419813Varnish cache server
Impossible to make sitemap too or make view with this content.... Problem limitation is still here and break all big website, please help us to find solution :)
Thx for all works Team.
- 🇫🇷France fgm Paris, France
The 503 in Varnish is not necessarily related to that issue: any page that is too long to serve will trigger it if your VCL has too short a timeout.
Can you reproduce the issue without Varnish ?
Patch no works with Drupal 10.1.5....
Patch oki D10.1.0 to 10.1.4.Still problem big content no works....
All Big site is break and bloque with Drupal 10....
- 🇫🇷France fgm Paris, France
Great to see this merged. However, unless I miss something it does not address to potential issue of saturating the static cache, as mentioned as part 📌 Add an entity iterator to load entities in chunks Needs work , does it ?
Do we want to track that along with the 2577417 iterator, or separately ?
I correct it on my website Drupal 10:
I up the SQL "max_allowed_packet" on my hosting to maximum 64MB and all is oki, it is configured at 1MB basically and too low for Drupal, best ways may be is to write in Tutorial to check this??? I read 8MB is a minimum for Drupal 9....