I'm trying to build a e-commerce site for a client and it isn't quite as straightforward as I had hoped. I am using Magento Community and my client has a wholesaler who provides a data feed (over 5000 products). In order to make this data feed compliant with Magento (and its attributes), I have edited some column headings in Excel and successfully uploaded as a CSV file.
My issue is that the wholesaler regularly renews the data feed automatically. When this happens, I am assuming my tweaking of the spreadsheet will be overruled, making my now Magento compatible CSV file useless again.
My question then is how can I make the wholesaler data feed compliant with my revised version so I don't have to continually rename elements? Is this possible?
I apologise if this sounds very stupid but I am more seasoned to static website builds.
Thank you in advance.
Does your webserver have access to the newly produced updated automatic data feed ?
IF so, just fetch the feed, modifty it to be complaint with the magento data feed then find the magento 'processfeed' or whatever function, and run it against the file ?
If your webserver cannot automatically fetch the file, can the pc making the feed automatically post the data to the webserver ?
IF neither are possible so far, then there will always be a manual person who has to 'reupload the feed'.. in which case, just make a simple page for them, 'Upload feed', or modify the magento page, and let them upload their standard feed, reformat it, then process the reformatted feed.
You would need to write a script to transform wholesaler data feed into magento format so you do not need to rename manually every time.
Related
I have created a tool using VBA that performs several pulls from BigQuery and does a variety of transformations and summarization inside of Excel. The tool is an evolving tool and to make sure that the users are using the most current version of the tool, I would like to have a version file saved on the shared Teams page that the tool would check and if needed, download the newest version onto the tool to the user's folder. I have scrubbed the web for several solutions to this ask but not one that actually works. Below are the links to the solutions I have attempted. The issue is that the connectors for opening text files seem to not work with a URL from Teams. I need to know how to make this work. Any help would be greatly appreciated.
https://www.excel-easy.com/vba/examples/read-data-from-text-file.html
This solution doesn't seem to work with a URL as I get a Bad Filename error.
https://chandoo.org/forum/threads/download-file-from-sharepoint-and-save-it-in-system-folder.41779/
Problem with this one is that the download for some reason shows completed but actually never moves local.
Connecting to Excel file stored on SharePoint via ADODB
https://www.connectionstrings.com/textfile/
I attempted to use this solution using the text file connection string but was not successful here either.
https://www.exceltip.com/import-and-export-in-vba/import-data-from-a-text-file-ado-using-vba-in-microsoft-excel.html
Again, another bad name error.
Any help or advice on how to connect to a simple text file shared in a Teams folder would be greatly appreciated.
So we need to migrate our old receipts from NetSuite filecabinet to another storage ( say AWS). How to do this without using any products like Celigo's SuiteStorage ?
Why not use an off-the-shelf product? It will get you results much quicker than trying to homebrew something, and you'll be able to move on to more important tasks, which should easily recoup any financial cost for a COTS product.
If your receipts are all underneath a single folder in the File Cabinet, you can just manually download the whole folder as a zip, then manually upload it wherever you want. Could take a while depending on how large the archive is, but it's a pretty simple process.
If you're hell-bent on building something yourself, there's no SuiteScript API for creating ZIP files, but you can SFTP individual files using SuiteScript 2.0 and the N/sftp module.
I'm hoping this is possible.
The organization I work for has a Sharepoint site and I am able to Upload Files to pages, however I am not an admin on our Sharepoint. I'm not sure what the version is, I think its older (ie: 2005).
I have some Excel Reports I've built. The data for these reports is pulled from a SQL Server Database which I have full control over. I have setup a Job in SQL Server to run every 12 minutes, this procedure pulls in some data and updates a few tables. These tables are used to feed my Excel Reports.
I have a separate Scheduled task set to open my excel report(s) refresh the data connections and save as a PDF.
I would like to link to these PDF Files via our Sharepoint so that the VIPs can access the reports as they want, but they always see the most up to date report.
I was trying to link to a Shortcut to the PDF Files but SharePoint doesn't seem to like that. How do I make the SharePoint link point to the PDF File that is saved over every 15 minutes?
Thanks in advance,
Any insight is greatly appreciated.
The way I do it (newish version of Sharepoint) is make the save location for the PDF the network location where Sharepoint keeps the files for that site. Usually you'll have access to those if you can edit the Sharepoint site.
Here is a tutorial to find that network location.
EDIT: It very well may be disabled by the admin at the moment. But it looks like the functionality is there.
Given the age of your SharePoint (either 03 or 07), most of the modern tools that you could use to do this don't exist for you (Excel reporting, BI tools, etc). The easiest solution I can think of is to actually modify the other side of the equation. A few options:
Change your report to output two copies of the same file. One entitled (as an example) currentreport.xls and the other report20150626.xls . Put the link to the currentreport.xls in SharePoint.
Build an ASP.net page that runs the SQL query you have built and pull the data through a view. Since this would be pulled on demand, it may be a few more cycles of your SQL code, but indexing, caching and selective data pull can prevent this from being an issue. Put the asp.net code in an iFrame in a SharePoint content editor web part.
Build your report using SSRS and host the output of that in SharePoint using an iFrame.
Run a scheduled job in SQL that copies your current report data to a table and query that table instead of your normal report table. That way you only have one Excel file that points to a specific table so no need to update links. You can always keep copying data to specific files if you need a historical record and can't use the DB to store this data for you (though the amount of space that it would take to do so would be minimal).
I am trying to download the full archive files of this website (http://www.afghanislamicpress.com/).
I tried using DeepVacuum (http://www.hexcat.com/deepvacuum/index.html) but the site is dynamic (I think that's the right word).
So you submit a form that gives the article archive, but it only spits out 5 at a time (i.e. per page) and then you have to click through. I want to download all the individual articles for the full data set, but don't want to manually click through.
I know there's some easy way to do this, but not entirely sure how.
Any suggestions for a novice at doing data scraping etc?
The most straightforward solution would be to contact the owner of the website and request their permission to republish their articles, and ask for a digital copy.
You can certainly automate pulling down content that is paged, but it requires some programming effort. The best tool for that imho is HTML Agility Pack.
Please be sure and comply with copyright and licensing terms of the content you are downloading.
I've been requested to help a friend where he wants all uploaded documents to be converted to PDF and save the URL into a small database for later use.
Can anyone explain me what should I look or what method I need to override to create such workflow?
This needs to be available to all documents (even images).
Thank you.
Have a look at this post, it does exactly what you need although you may need to pull some tricks to write the URL into an external database. Writing it to a SharePoint list is easy.
Ping me if you need help.
It would be better to use third party API like one suggested by Muhimbi. to create your custom one you would have to create a event handler in which you can write code to convert your document or images to PDF. for PDF conversion you can use itextsharp
http://sourceforge.net/projects/itextsharp/