I have been indexing for quite a long time but now I am not able to do so. I keep on getting the following error.
INFO [Thread-80] (00002SB6) [SolrIndexerJob] Started indexer cronjob.
ERROR [Thread-80] (00002SB6) [Job] Caught throwable de/hybris/platform/solrfacetsearch/config/IndexConfig
java.lang.NoClassDefFoundError: de/hybris/platform/solrfacetsearch/config/IndexConfig
at ma.glasnost.orika.generated.Orika_FacetSearchConfig_FacetSearchConfig_Mapper45623933135018$4.mapAtoB(Orika_FacetSearchConfig_FacetSearchConfig_Mapper45623
933135018$4.java)
at ma.glasnost.orika.impl.mapping.strategy.UseCustomMapperStrategy.map(UseCustomMapperStrategy.java:67)
at ma.glasnost.orika.impl.MapperFacadeImpl.map(MapperFacadeImpl.java:735)
at ma.glasnost.orika.impl.MapperFacadeImpl.map(MapperFacadeImpl.java:714)
at ma.glasnost.orika.impl.ConfigurableMapper.map(ConfigurableMapper.java:150)
at de.hybris.platform.solrfacetsearch.config.impl.DefaultFacetSearchConfigService.getConfiguration(DefaultFacetSearchConfigService.java:51)
at de.hybris.platform.solrfacetsearch.indexer.cron.AbstractIndexerJob.getFacetSearchConfig(AbstractIndexerJob.java:70)
at de.hybris.platform.solrfacetsearch.indexer.cron.SolrIndexerJob.performIndexingJob(SolrIndexerJob.java:49)
at de.hybris.platform.multicountry.solr.indexer.cron.impl.MulticountrySolrIndexerJob.performIndexingJob(MulticountrySolrIndexerJob.java:73)
at de.hybris.platform.solrfacetsearch.indexer.cron.AbstractIndexerJob.perform(AbstractIndexerJob.java:40)
at de.hybris.platform.servicelayer.internal.jalo.ServicelayerJob.performCronJob(ServicelayerJob.java:38)
at de.hybris.platform.cronjob.jalo.Job.execute(Job.java:1390)
at de.hybris.platform.cronjob.jalo.Job.performImpl(Job.java:814)
at de.hybris.platform.cronjob.jalo.Job.performImpl(Job.java:732)
at de.hybris.platform.cronjob.jalo.Job.perform(Job.java:644)
at de.hybris.platform.servicelayer.cronjob.impl.DefaultCronJobService.performCronJob(DefaultCronJobService.java:86)
at de.hybris.platform.solrfacetsearchbackoffice.wizards.BaseSolrIndexerWizardStep$WizardCronJobAsyncOperation.execute(BaseSolrIndexerWizardStep.java:158)
at com.hybris.cockpitng.engine.impl.DefaultWidgetInstanceManager$1.getResult(DefaultWidgetInstanceManager.java:206)
at com.hybris.cockpitng.engine.operations.ResultLongOperation.execute(ResultLongOperation.java:52)
at com.hybris.cockpitng.engine.operations.LongOperation.run(LongOperation.java:205)
at java.lang.Thread.run(Thread.java:748)
Not sure which change is causing this issue.
Tried to setup new fresh hybris suite still the same issue. Followed hybris answers which mentioned to perform ant clean all and server startup, didn't work. Created new index and then performed indexing still the same. I am able to open solr admin but not able to index anything.
Any help would be really appreciated.
Just restarted the server. It worked for me!
Related
I'm responsible for upgrading our Lucene indexes from 4.5.1 to 8.x. Thanks to the unofficial Lucene Upgrader I was able to do this, so now I'm testing this index in Solr 8.11.2.
When I launch the standard *:* search through the Admin interface, the return message told me about a NullPointerException:
java.lang.NullPointerException
at org.apache.solr.legacy.LegacyNumericUtils.prefixCodedToLong(LegacyNumericUtils.java:191)
at org.apache.solr.schema.TrieField.toObject(TrieField.java:154)
at org.apache.solr.schema.TrieDateField.toObject(TrieDateField.java:94)
at org.apache.solr.schema.TrieDateField.toObject(TrieDateField.java:86)
at org.apache.solr.schema.TrieField.write(TrieField.java:280)
at org.apache.solr.response.TextResponseWriter.writeVal(TextResponseWriter.java:125)
At first, I had no idea how to debug this, and googling it didn't return anything relevant. Then I got the idea of commenting the date fields in schema.xml. This still got me a NPE, but now it comes from a regular TrieField:
java.lang.NullPointerException
at org.apache.solr.legacy.LegacyNumericUtils.prefixCodedToLong(LegacyNumericUtils.java:191)
at org.apache.solr.schema.TrieField.toObject(TrieField.java:150)
at org.apache.solr.schema.TrieField.write(TrieField.java:280)
at org.apache.solr.response.TextResponseWriter.writeVal(TextResponseWriter.java:125)
Surely there's a better way to debug this than to comment every field? Or is there something I'm missing?
P.S.: Reindexing is not an option. We have several customers where indexing literally takes weeks, if not months.
I've been trying to set up CKAN, however I am facing some problems in SOLR.
Everytime I run CKAN, the SOLR log file registers a new event, and that is:
org.apache.solr.common.SolrException: sort param field can't be found: metadata_modified
I am trying to use CKAN for the first time and I have no experience at all, so I have no idea what that log event means, nor how to fix it.
EDIT:
When I reload the core at SOLR, the following is logged:
The schema has been upgraded to managed, but the non-managed schema schema.xml is still loadable. PLEASE REMOVE THIS FILE.
Could anyone help me?
Many thanks.
Well, turns out the SOLR 6, for some reason, was the problem.
Downgrading to version 5 worked for me.
This sounds like you are not using CKAN's custom Solr schema. Make sure to go over all points in the setup documentation, specially point 2, and to restart jetty afterwards:
sudo service restart jetty
Solr 4.10.4 wont allow me to create a core.
When I try creating a core I get the following error
Error CREATEing SolrCore 'blog': Unable to create core [blog] Caused by:
Can't find resource 'solrconfig.xml' in classpath or
'/Users/ra/desktop/solrx/solr-4.10.4/example/solr/blog/conf
I'm using the admin to do this. I also tried
bin/solr create -c blog
and it still doesnt work how do you create a core in solr 4.10.4
I figured it out. the tutorial I am reading says to place blog in
example/
when it's supposed to be in
example/solr/
tht made it work hope it helps someone
After moving from Solr 4.4 to Solr 4.6.1, I am getting the below Exception while updating my Indexes using the Data Import Handler. Does anybody have any ideas on why this is happening?
ERROR - 2014-02-18 09:39:35.232; org.apache.solr.update.StreamingSolrServers$1; error
org.apache.solr.common.SolrException: Bad Request
request: http://10.200.131.174:8080/solr/collection1/update?update.distrib=FROMLEADER&distrib.from=http%3A%2F%2F10.200.131.173%3A8080%2Fsolr%2Fcollection1%2F&wt=javabin&version=2
at org.apache.solr.client.solrj.impl.ConcurrentUpdateSolrServer$Runner.run(ConcurrentUpdateSolrServer.java:240)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
ERROR - 2014-02-18 09:39:35.244; org.apache.solr.update.StreamingSolrServers$1; error
org.apache.solr.common.SolrException: Bad Request
This is a known issue for Solr versions greater than 4.5. I too ran into this while running SolrCloud with 2 shards. It seems that the update works fine on the server running DIH, but fails when it is forwarded to another shard or replica.
Here's the related jira issue. And here's another unlucky soul like you and I who also faced this error.
As far as I know, the only workaround currently is to go back to a Solr version 4.5 or below...
Am using Apache Solr 3.1.0 with Lucene 3.1.0. While indexing server got stopped due to hardware failure. I restarted my server but solr indexing is not happening.
am getting
SEVERE: java.lang.RuntimeException: org.apache.lucene.index.CorruptIndexExceptio n: document count mismatch: deleted docs count 0 vs segment doc count 148 segmen t=_otm
It seems to be a bug in below 3.6 versions reference. But how to upgrade to higher version with out losing the old indexes.
How to solve this. Any body can tell me.
Upgrading won't fix your problem.
Have a look at CheckIndex class. You can first run checkIndex() method (read-only) which would provide with the status information (CheckIndex.Status) and then you can run fixIndex(CheckIndex.Status status) in order to attempt to fix it. Good luck!