Does anyone know if it is possible to share data between sites on a Kentico platform, such as news articles? I have tried searching but could not find an answer.
Thank you,
Yes. If you want the content to be the same and keep the URLs and Domains separate, then this is best accomplished with linked documents between the sites. This way if they are updated on either site, they will remain in sync on the other site. Read more here: https://docs.kentico.com/k9/managing-website-content/working-with-pages/copying-and-moving-pages-creating-linked-pages
https://docs.kentico.com/k9/e-commerce-features/managing-your-store/products/linking-existing-products-or-sections
Basically what you'd want to do in your content tree, is click "new page" and then select "Link an existing page" from the bottom of the next menu. After that, you will see a content tree to select a page from. Use the site selector in the top left to choose a different site, and then select the page you'd like to link in.
Note: Keep in mind that the page type will need to be allowed in that section of the tree (for example, if you are trying to nest a news article under a folder but folders are not an allowed parent type, you will get an error).
If you'd like to pull data from another site in via repeater and leave the content on the other site, then you can user a repeater or similar viewer control, and specify a site within the properties in order to pull the pages from the other site.
Each of these methods assumes the Kentico sites are on the same instance of Kentico with a shared database.
If you have multiple sites in one Kentico instance, in your repeater getting the news articles, specify the site you want to get the news articles from and it will be displayed on that site. If you want to combine them then create a custom query and use the query repeater to show the news articles.
Related
I am currently involved in project where we are using Liferay (6.1 GA2).
It seems that Liferay search results provide links to Web Content Fragments instead of to the pages containing them.
Have any of you gone through this issue? Do you know how to solve it?
Thanks a lot pals.
Best, Alberto
You can have a lot more content in the backend than actually displayed on any page. Further, you can display any article on multiple pages at once.
A way to work around this is to specify in the "Web Content Search" portlet that you're only interested in content that is actually published. However, this does not solve your second problem: The content can still be published on many different pages.
Every content can have a "Display Page" - the setup of such a display page is well explained in the UI (see the Web Content Editor) so that you'll actually see a proper page with the search results.
If you actually want to search for pages only instead of content (you might miss out on some metadata), I'd recommend to go with some spider solution that spiders your website, indexes the pages independent of their construction elements (articles) and search that external index.
I want to find 3 top most viewed pages in the entire site collection and display their page titles. Using Web analytics webpart, I got top most 10 pages URL. But i want the page title and the not URL. Is there anyway to do that?
Regards,
Raji
This post(http://auditlogsp.codeplex.com/) can give you a reference. The solution can help us to log the source of SharePoint events , actually include the 'Viewing' action.You can store those information at an isolated database for analyzing the “top” reports.
Sharepoint 07 has nice page library. You can create templates and then just by pressing Site-> 'Create a page' get a choice with page templates.
However if one tries to create second library, for example, document library, in the same site (node), then it is not possible to choose element type as "page". Only "wiki-pages" and "pages with web-parts"!
So, is it possible to get 2 or more page libraries in the same node?
It seems very inconvenient to create the hole new site or new node only just for page library.
The SharePoint Publishing Infrastructure can only handle a single Pages library per Web site, it's a physical limitation of the software.
You could create sub webs to separate out your content more, or create new Content Types inheriting from the "Page" content type, and add in extra columns to store metadata to further categorize your pages. You'd then add this new content type to the Pages library, and you can slot your new content into this content type to categorize it.
Luckily, I have discovered another solution. Using Metalogix Content Matrix I was able to copy and paste the 'Page' Library renaming it and Viola...it works! I'm able to edit pages without getting the "The SPListItem provided is not compatible with a Publishing Page" Error. Whohoo! I hope this helps you.
I am using Microsoft SharePoint Search (MOSS) to search all pages on a website.
My problem is that when you search for a word that appears in the header, footer, menu or tag cloud section of the website, that word will appear on every page, so the search server will bring you a list of results for that search term: every page on the website.
Ideally I want to tell the search server to ignore certain HTML sections in its search index.
This website seems to describe my problem, and a guy says "why not hide those sections of your website if the User Agent is the search server.
The problem with that approach is that most of the sections I hide contain links to other pages (menu's and tag clouds) and so the crawler will hit a dead end and won't crawl very far.
Anyone got any suggestions on how to solve this problem?
I'm not sure if i'm reading this correctly. You DON'T want Search to include parts of your site in the index, but you DO want it to go into that section and follow any links in it?
I think the best way is to indeed exclude those section based on user agent (i.e. add them to a usercontrol and if the user agent is MS Search you don't render the section).
Seeing as these sections would be the same on every page, it's okay to exclude them when the search crawler comes by.
Just create ONE page (i.e. a sitemap :-D). that does include all the links a normal user would see in the footer / header / etc. The crawler could then use that page to follow links deeper into your site. This would be a performance boost as well, seeing as the crawler only encounters the links once instead of on every page.
Can someone give me some directions on how to setup SharePoint Search Center so I can get results from the list and that they have some custom (modified) link?
I have Forms authentication (and anonymous access) enabled with alternate access mapping.
Right now in the Default zone I get results from the data in lists and they all point to the AllItems.aspx. If try search from the Internet zone I don't get any results from the lists and I am guessing that this is because of some security settings. But if make them to show how will I customize resulting link so that list items are shown with some publishing page.
For example if I keep news in the News list and when I do search I want to get result with link in following format
http://somesite/Pages/News.aspx?itemId=12
where the itemID is he id of the news item.
Can I customize link in the result ?
You can customize the result link using the Core Search Results web part. It is all in the XSL which is available if you modify the shared properties of the web part.
The problem is that this page is meant to show search results of all types including documents in SharePoint, files potentially outside of SharePoint, web pages, business data, etc.
You may want to have a custom search results page that uses a specific scope or managed property query such that you can be sure the results will be list items. This can probably be done without any coding (if you don't consider XSL coding) and you could still use the Core Search Results web part.
Another option may be similar, but use the Data Form/View web part (through SharePoint Designer) or the Content Query Web Part (Publishing Infrastructure feature required).