Jobs and schedules - how do they work with each other? - cognos

could please someone enlighten me?
I have a data set and I need this data set to be refreshed on the 5th and 15th of the month.
But there can only be one schedule per Cognos object, right?
Is this the optimal solution?
I setup a schedule on the data set itself for the 5th of Month.
Then I create a job and add that data set as a job step.
I schedule the job containing that data set (which includes its own schedule) for the 15th.
What happens on the 15th and how these two schedules work with each other?
(This could also apply to reports etc).
Many thanks
I went through the documentation and some online articles, but there is no definitive answer if a data set (or a report..) can have multiple schedules set.

Lots here about reports and report views explaining how I handle schedules. Substitute "dataset" and "job" and I think you'll get what you want.
One schedule per object leads to limitations. I have found for reports that it is better to create a report view to handle the schedule. I very rarely schedule a report anymore because I often end up needing more than one schedule or set of parameters.
Using a report view provides a means to not only create multiple schedules per report, by creating multiple report views, but also have those schedules send the reports based on independent sets of parameters. For example, the regional VPs want reports showing their sales revenue for the month, but only for their regions. I would create one report, then create a report view for the U.S. and a report view for Europe. Both get sent on the 5th of the month.
Likewise, I could set up 2 report views to work around the scheduling limitations. Say I want to send the same report on the 5th and the 15th. I would create the report, then create 2 report views to handle the schedules.
So, to translate this to your need...
To refresh the same dataset on the 5th and the 15th, create 2 jobs that have the dataset as a step -- one job for each schedule.
Using jobs looks like the right thing to do for your use case. I don't see another scheduling option for datasets.

Related

Is there a way to use an IF statement to track if another user changed the data in a table and time stamp it?

I have an access database in which multiple customers will be placing large orders on a weekly basis. I have a query automatically set up in excel to refresh the information every minute. A potential issue I could see happening: Manager places order in our access form, decides to edit his record later w/o contacting the scheduler, and changes the order information he originally entered. I would like to track what it was changed from and to, as well as the date they changed it. These orders have to be placed by a certain time in the week for them to make the shipment in or out the following Monday. If orders are changed, I want to know. If orders are changed after the deadline, I especially want to know. Thanks
Create a before update event that inserts the changes into an audit log table. Here is the resource I used for my databases: https://www.techrepublic.com/article/a-simple-solution-for-tracking-changes-to-access-data/
Note: This will only track changes made to existing records through the form. It also will not track deletions or changes made directly in the table.

Data import from new reports and automatically update existing records in existing data

I seek expert advice of all of you in accomplishing my work related task.
Task : Task it to perform Analysis on reports obtained to WorkSafe Monthly and weekly and getting valuable information out.
For example :
Number of injuries on monthly basis drilled down by department and divisions.
Total days lost in year
Count of type of claims
Possible return date.
So i receive these reports and i add some modified columns to it. Like Correct employee names and their ID's just to create a relationship between Employee database in Powerpivot so that i could get their position, dept and division.
Now every month in new report there could be 2 or 3 new claims added to it, and some existing claims with updates info. like Updated return to work date, Short term disability days etc.
Currently i go through them manually and it's really time consuming and tiring. if there there was older claims weren't getting updates i could've just imported from a folder using power query and added steps to remove duplicates. However, if i remove duplicate claims using powerquery now, basically i'll be removing same claims with updated info.
Could anyone you here suggest an efficient way to do it with power query or with other approach?
Thanks in Advance. I'd appreciate your time and effort.
If you use PowerQuery and select your source From File -> From Folder, when you choose to combine and edit you will get a table with its first column named Source.Name; which you could use to differentiate the updates.
For instance: If I start with two excel files in the same folder (theoretically, they could be different date source files for you)...
WS1.xlsx:
WS1 - Copy.xlsx:
Then I use the folder they are both in as the source...
(Navigate to your folder as appropriate.)
...and select Combine & Edit...
...and select the worksheet...
Then I get this:
...and it is clear what information came from what source file.

SharePoint Office 365 - Starting the ID field at 50,000 which incrementally grows

I would like to premise that I am not a coder, and have limited SharePoint experience. Nonetheless, I have a task at work to create a new tool for a group at work. This tool will allow them to track things through its life-cycle. The old tool will stop being used once they hit the 49,999th record. At that point we will transition to SharePoint. The prior numbers will live with those records and we will be unable to start back at the number 1.
My goal is to have an ID field for each record that incrementally grows starting from the number 50,000.
I tried creating a new calculated formula field that does 50,000 plus the ID but it does not work effectively. There tends to be a lag. The lag can be anywhere from 5 minutes to a day or two. In turn, I need an alternative option.
With a friends coaching, I am messing with SharePoint designer to automatically create a record once one is deleted and delete one every time one is created. However, this times out occasionally. Even more so, it is extremely slow. It will take me a week or so to get to 50,000. Does anyone have any experience or expertise to assist me? It is very appreciated.
The way I did it was to create another field and used that in all my views. Update it through a SharePoint Designer workflow on create - set the field to ID+50000, for example, or whatever you need. Maybe give it a bit of a buffer and set it to ID+51000. By the time the workflow is running the ID has been created and you can use it in a workflow.
If you are going to be moving your old records to the list, you can change the workflow to only update the new field if it's blank. That way you can save the original ids.

SharePoint List - Daily email even when there are no changes?

(As far as I know) SharePoint list alerts only fire off due to a change. If I want an email to go out daily even if there were no changes, I figured I could create a calculated field that does today's date minus the "Start Date". It will increment by 1 every day. Will this cause a daily change email to go out? If not, does anyone have a trick for emailing out a list daily even when there are no changes? Thanks!
This will not cause the alert to go out. Calculated fields are only calculated when the list is accessed. The only way I see to do this is to create a site level workflow with a tool like Nintex or create a timer job that can generate the email you want. Nintex Workflows can be scheduled to run on a regular basis. Workflows created with SharePoint Designer cannot be scheduled out of the box.

Business Objects Web Intelligence automate report once for each parameter

Using Business Objects Web Intelligence or Rich Client, is it possible to schedule a report to automatically run and export to Excel once for each parameter?
Background: I have a single report that needs to be saved to different Excel files once a month but I have a parameter on a "client" dimension which can contain 30+ values. Each value needs to generate its own Excel file. I would prefer to not run the report 30+ times each month.
Is there a way to schedule or change the WID query so that a separate file is generated for each parameter value? Or maybe an alternative idea for this manual process?
This can be done using a Publication with Dynamic Recipients.
You would need to create a second report that would produce a list of all values in your Client dimension. This report would then serve as the source of the Dynamic Recipients in the Publication. In the base report, you would remove the prompt on Client but make sure that Client is included in the report results. Then in the Publication, you would set Personalization such that a filter is automatically applied to the Client dimension based on the values from the second report.

Resources