Can I continue solr indexing if an error occurs due to bad data instead of abort - solr

We are using Primero and it uses Solr for search.
Below is the command to reindex
RAILS_ENV=production bundle exec rake sunspot:reindex --trace
However, we get an error due to bad data that is hard to find. Is it possible to not abort and continue reindexing on error?
Error message
rake aborted!
ArgumentError: invalid date
Also, is it possible to get a detailed error?
Any help is appreciated.

The problem was with the error handling in sunspot package.
There was a bad date entry empty space for one of the records that failed when Date.parse() tried to parse the date.
I had to debug the sunspot code to figure out the field that resulted in this issue. After correcting the record reindexing worked.

Related

Search for solve a error 255 in SOLR Nutch

I have allready tested SOLR with Nutch on local machine and it runs without Errors. Now I have deploy it to live server and got this error 255
enter code here
Indexing 139/139 documents
Deleting 0 documents
Indexing job did not succeed, job status:FAILED, reason: NA
Indexer: java.lang.RuntimeException: Indexing job did not succeed, job status:FAILED, reason: NA
at org.apache.nutch.indexer.IndexingJob.index(IndexingJob.java:150)
at org.apache.nutch.indexer.IndexingJob.run(IndexingJob.java:291)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
at org.apache.nutch.indexer.IndexingJob.main(IndexingJob.java:300)
Error running:
/home/apache-nutch-1.18/runtime/local/bin/nutch index -Dsolr.server.url=http://localhost:8984/solr/nutch -Dmapreduce.job.reduces=2 -Dmapreduce.reduce.speculative=false -Dmapreduce.map.speculative=false -Dmapreduce.map.output.compress=true crawl/crawldb -linkdb crawl/linkdb crawl/segments/20220728170007 -deleteGone
enter code here
Failed with exit value 255.
Anyone an idea what this can be?
Show if the core exists in SOLR
install ant if not present
go to $NUTCH_HOME and run > apt runtime
Thats all folks and will now run and write

GAE 1.7.3 - How to debug appengine-web.xml XML error validating

How do you fix xml validation errors in appengine.web.xml.
First the error message told me to add "true".
And now it just fails, saying that the valiadtion failed - but, what to do?
I did fix this, by creating a new ApEngine project and copied it over, but how do you find the errors?
Looking for the process on finding the errors?
Regards

How to upgrade cakephp from 1.3 to 2.1?

I'm a newbie in cakephp, I'm trying to upgrade cakephp to the latest version.
I install the fresh cakephp 1.3 on my computer and the upgrade it to cakephp 2.1.
I use shell to upgrade, but after I run 'upgrade all' command, I saw two error:
Warning Error: chmod(): Operation not permitted in [/var/www/cakephp-1.3/lib/Cake/Utility/Folder.php, line 639]
Warning Error: touch(): Utime failed: Operation not permitted in [/var/www/cakephp-1.3/lib/Cake/Utility/Folder.php, line 640]
I think it has upgraded complete. Because I see the message from terminal like this:
Done updating /var/www/cakephp-1.3/app/Console/cake.php
Done updating /var/www/cakephp-1.3/app/Console/Command/AppShell.php
Running components
Running exceptions
Then I refresh my app and I got some errors:
http://flic.kr/p/bwUpwY
Then I delete 'cake' directory, and the error message has changed:
http://flic.kr/p/bKP7Te
So now I don't know what to do next, because I did many ways but still not make it work.
So anybody please tell me what I did wrong and how can I upgrade cakephp successful.
Thanks in advance.
This looks like you have permissions issues on some of the directories that the CakePHP migration script expects to be writable.
This could lead to the migration failing on some parts and leaving a partially broken install after it completes.

org.apache.solr.common.SolrException: missing content stream

I have installed Apache Solr with Tomcat and my /solr/admin is working fine. But when I try to issue /solr/update I am getting the following error. What could be the reason?
org.apache.solr.common.SolrException: missing content stream
If you add commit parameter i.e. ?commit=true, it will work
/solr/update will look for any input documents to be indexed. Running plain /solr/update will cause this exception since there is no input for it. The easiest way to run it is like,
java -Durl=localhost:8080/<your apache solr context path, mostly solr>/update -jar post.jar *.xml
This can also happen through SolrJ/spring-data-solr if you try to persist an empty collection of documents.
So solrClient.add(new ArrayList<SolrInputDocument>(), 10000);
would also cause the error.

Google App Engine - Error vacuuming indexes

I had an index that had status error, so following the docs I tried to vacuum my indexes to remove it. When running appcfg.py to vacuum it I got this error message:
Error 400: --- begin server output ---
Deleting a composite index failed:
ApplicationError: 1
--- end server output ---
Now all of my indexes are in error state. How can I fix this?
Must just have been a temporary issue, I retried the exact same command today and it worked.

Resources