Stream: troubleshooting

Topic: โœ” Draft version of published datasets stays around


view this post on Zulip Henning Timm (Apr 24 2024 at 12:57):

I stumbled across another problem :sweat_smile: We have a dataset that has been published, but the interface is still showing an unpublished draft version (only). Clicking on the dataset forwards me to the published version v1.0 of the dataset, since the draft version is no longer around. Things I tried:

Things I have not tried:

I have seen some JVM crashes during SOLR runs that occurred yesterday. I am not sure if they are related. This might be related to our problems with full text indexing that we had earlier.

Has anyone encountered something like this before?

view this post on Zulip Philip Durbin ๐Ÿš€ (Apr 24 2024 at 13:23):

Huh. I would expect that reindexing the specific dataset should remove the draft version. It's strange that it did not.

You could delete the document from Solr by calling Solr directly.

It sounds like you tried https://guides.dataverse.org/en/6.2/admin/solr-search-index.html#manually-querying-solr already. Inconclusive? Did it return two Solr documents? One published one draft?

view this post on Zulip Henning Timm (Apr 24 2024 at 13:44):

Thanks for the help! The first problem is solved, the manual query returned no results because I messed up the query URL. Now it shows all datasets.

For the dataset in question, SOLR only return the draft version, but it seems like all SOLR reindexing runs crash with an OutOfMemory Exception. I think I need to dig into my settings here.

view this post on Zulip Philip Durbin ๐Ÿš€ (Apr 24 2024 at 13:48):

Are you saying you get out of memory problems even if you try to reindex a single dataset? :grimacing:

view this post on Zulip Henning Timm (Apr 24 2024 at 14:00):

Correct :sweat_smile: I re-index the single dataset and get a nice jvm crash report

# A fatal error has been detected by the Java Runtime Environment:
#
#  Internal Error (debug.cpp:362), pid=3178922, tid=3178961
#  fatal error: OutOfMemory encountered: Java heap space
[...]
Command Line: -Xms512m -Xmx512m -XX:+UseG1GC -XX:+PerfDisableSharedMem -XX:+ParallelRefProcEnabled [...]

I am not this experienced with SOLR. Are the 512m a reasonable value?

view this post on Zulip Henning Timm (Apr 24 2024 at 14:15):

It seems to work with the same settings on other servers, so I am not sure what is going on here. I will try a complete re-index later today.

view this post on Zulip Philip Durbin ๐Ÿš€ (Apr 24 2024 at 14:51):

Hmm, I'm not sure.

I do see 1 GB here: ExecStart = /usr/local/solr/solr-9.3.0/bin/solr start -m 1g

Which is twice what you have. That's from the solr.service script at https://guides.dataverse.org/en/6.2/installation/prerequisites.html#solr-init-script

view this post on Zulip Philip Durbin ๐Ÿš€ (Apr 24 2024 at 14:52):

Perhaps @Don Sizemore or @Leo Andreev can suggest a sufficient amount of memory for Solr.

view this post on Zulip Philip Durbin ๐Ÿš€ (Apr 24 2024 at 14:53):

But obviously, you shouldn't get a crash like that when reindexing a single dataset. :scream:

view this post on Zulip Don Sizemore (Apr 24 2024 at 15:08):

Henning Timm said:

Correct :sweat_smile: I re-index the single dataset and get a nice jvm crash report

# A fatal error has been detected by the Java Runtime Environment:
#
#  Internal Error (debug.cpp:362), pid=3178922, tid=3178961
#  fatal error: OutOfMemory encountered: Java heap space
[...]
Command Line: -Xms512m -Xmx512m -XX:+UseG1GC -XX:+PerfDisableSharedMem -XX:+ParallelRefProcEnabled [...]

I am not this experienced with SOLR. Are the 512m a reasonable value?

Phil is correct; Harvard had trouble with Solr stability and bumped their SOLR JVM heap size to 1GB in response.

view this post on Zulip Henning Timm (Apr 24 2024 at 15:14):

Don Sizemore said:

Henning Timm said:

Correct :sweat_smile: I re-index the single dataset and get a nice jvm crash report

# A fatal error has been detected by the Java Runtime Environment:
#
#  Internal Error (debug.cpp:362), pid=3178922, tid=3178961
#  fatal error: OutOfMemory encountered: Java heap space
[...]
Command Line: -Xms512m -Xmx512m -XX:+UseG1GC -XX:+PerfDisableSharedMem -XX:+ParallelRefProcEnabled [...]

I am not this experienced with SOLR. Are the 512m a reasonable value?

Phil is correct; Harvard had trouble with Solr stability and bumped their SOLR JVM heap size to 1GB in response.

Thanks! I'll give this a try.

view this post on Zulip Henning Timm (Apr 24 2024 at 16:55):

That did the trick! Thank you for the help!

view this post on Zulip Notification Bot (Apr 24 2024 at 16:55):

Henning Timm has marked this topic as resolved.


Last updated: Oct 30 2025 at 06:21 UTC