What is the best approach for using SOLR with web projects?

ok, I am completely new to SOLR and Lucene, but I have Solr running out of the box under Tomcat 6.x and just moved on to some Wiki elements.

I have a few questions and you also need some suggestions.

  • Solr can index data in files (XML, CSV), as well as index DB. Can you just point it to the URI / domain and index the website so that google?

  • If I have a site with the data "Pages", so "Page name", "Page content", etc. and "Product Data", so "Product Name", "SKU", etc., do I need two different Schema.xml files? and if so, does that mean two different Solr instances?

Finally, if you have a project with a large relational and normalized database, what would you say is the best approach of the 3 following options:

  • You have a background middleware service that runs the database and manually creates the appropriate XML files to send to SOLR

  • Have a SOLR index directly in the database. In this case, would it be best to point SOLR to views that abstract all table relationships?

  • Any other options that I don't know about?

Context: we are running Windows 2003, .NET 3.5, SQL Server 2005/2008

Hurrah!

+5
source share
2 answers
  • No, you need a scanner for this, for example. Nutch
  • , (= schema.xml), , , . Solr, Cores.

Solr, , , .

:

  • - ORM, . , NHibernate , . NHibernate SolrNet,
+7

, . , , , . ( ?) , Solr, DIH . , , , , Solr, .

+1

All Articles