My company use a hosted web server, so I canĀ“t install custom search engines based on java like Lucene-search. Are there another way to display the search results without markup?
There are a zillion hosted elasticsearch providers; it doesn't have to be installed on the same server than MediaWiki.
Related
We are currently using a number of open source and commercial products to store different type of information (in our internal network). All these products come with their own repositories (usually a database) and their own search capabilities and store different type of information.
Currently the list of products is as follows:
Wordpress
Jira
Confluence
Sharepoint
Dynamics AX
Moodle
The problem we are facing is that when one needs to search for information, one needs to login into all these different systems and execute a search on each one.
I Googled for "search engine frontend", "meta search engine", etc. but i was not able to find something obvious that solves our problem. At this point, i have to say that we are not interested in building one "central repository" to be searched, but instead we are in need of a frontend that will accept the query from the user, "package it" to the format that each of the individual search engines understand, receive the respone (JSON or XML) and present it to the user
Any suggestions on how we could solve it?
Your strategy is right: If you are not interested in building a central index, you will need an application that accepts the query from the user, converts it to the format that each of the individual search engines understand, receives the responses and presents them to the user. This is exactly what a meta search engine does. Even if you use a framework (e.g. Carrot2), much work will probably remain to write those query and result transformers, and you will probably experience slow results because the meta search can never be faster than the underlying search modules of the components you search through.
Instead of querying each backend separately you can put your data into one backend.
You could export your data to a Apache Solr server and use a frontend like CorePages, http://www.corepages.biz . You could add a backlink to your data so you can directly jump to your search result entry, f. e. a Jira Ticket or a wiki article.
We have built a web site which employs Lucene.NET for search. We recently have integrated another web site so that form a user's perspective both website seem to be just one site! (we share the mater pages, etc.)
The problem we have is that two web sites are hosted in different locations. So when Lucene.NET crawls the first web site, it does not pick the content of the second web site. We want to extract the content from the 2nd web site and put it in the same index file that is built for the first site.
How can I get Lucene.NET to crawl an external site too?
Thanks
If you have file-system access to the 2nd system than you can just index by providing the path. If not, you will need to write a crawler, you can start with something basic using HttpWebRequest, or get fancier by using some tools that recursively crawl a site using links etc.
I am trying to make a kind of 'faceted search' in ASP.NET and am looking for good ways to filter data. That's why I have the following question:
Is it possible to use FAST search on a SQL database using Microsoft Search Server 2010 without using SharePoint? (in other words: can a SQL DB be a content source for the search?) I can't seem to find a definitive answer to this anywhere.
I have found this on the MS website:
The following connectors come out of the box in Search Server 2010
Express and SharePoint Server 2010:
SharePoint sites
Windows file shares
Exchange public folders
Lotus Notes
Web sites
IFilters for additional repositories
Structured content in databases
Federation object model
I would like to be sure that it 'can' work before I start messing around with it.
You will need to install SharePoint Foundation (free version), as well as install the cumulative updates/service packs to install and configure the Secure Store Service. From there, you can use Business Connectivity Services to model your database and present it to Search Server Express. FAST Search is another product altogether and would require additional software installation and configuration.
i have a site that im developing and i want to put there content and links (some of them are my private home links and stuff) drupal search is good, but its not accurate and also its shows me only things from my site... i want to search all include my website, but dont let google to public it world wide... is taht possible?
It's not possible to have google index your site without making it accessable to the web. However you can set up a local solr server and user the Apache Solr Search Integration Module to index your site, which is much better than the normal drupal search. That would only provide search for your site still. I'm not sure what you mean by "only shows me things from my site."
If you are not able to host a solr server an alternative is Acquia's hosted solr search. if the pricing is right for you it's a quick and easy way to get solr search on your site.
For some reason my search in the sharepoint site does not work.
I have set up the SSP, the scopes, the crawls, everything but it still does not work
Can someone explain to me how to setup the search? Maybe I did something wrong in the process.
It's not the simplest thing in the world to setup, as it's comprised of a number of components.
You need to check each one to determine where your problem is.
Start from the crawl, and work your way forward to the search production on the page.
So check the following:
Check some servers have been setup to index pages. (You can see this under services on servers in the central administration pages.)
Make sure they're all running correctly. (Not in a half started state.)
Check your crawl log in your SSP to see if it is indexing anything.
(Index different types of content, like file shares, web sites, and sharepoint itself. (check each one.)).
(Note you need a special plugin to index PDF's.).
Check your index is copied to the front end server where it is used.
If it's not, it may be because this hasn't been configured, (Check Services running on servers again)
Then check your site collection setup, and ensure you have a search site configured.
Ensure the site collection search details are configured to use the search site.
Finally check the user doing the searching actually has access to the content being indexed.
Doing all of that should give you some idea of where the problem is.
In addition to Bravax's answer its worth checking that you are not getting stung by the local loopback check.
I had similar problem and ended up using search server express which is free (see my answer from this link: sharepoint 2010 foundation search not working)
I have installed search server express 2010 on top of SPF which works great. it has additional features and work well with sharepoint foundation. her is a link for upgrade and configuration: http://www.mssharepointtips.com/tip.asp?id=1086
You need to crawl the the contents source and add the website to it, then run full crawl to index data.