So, I'm going along in SSRS building a new report, as you do, and as I'm building the query from my SSAS cube I get the error: "The Cubes collection has been updated on the server." I can't find anything on the interwebs that would fix my issue. The odd thing is that I can build out the query for a while before getting this error, meaning I'm able to add 7-10 fields before the error presents itself. After the error I can't save it because it can't verify that the query is working. So I have to close it all down and restart building the query, but again, after a few fields get added I get the same error.
Yesterday we added a tabular instance to our BI server so we could convert our SSAS cubes to a tabular model and then utilize power bi and this seems to be when I started to see the issue. I'm not sure if they're related. I did restart the server last night hoping that would correct anything that went awry, but to no avail.
Cubes need processing. Process your cubes and try again.
Related
I am having a strange issue with SSRS 2016. We setup a new server and uploaded some reports, the reports ran fine for few days. The issue is, randomly the report doesnt run, When I load the report page, the parameters appear blank and I am unable to select any parameters. If I try to run and click the view reports button, nothing happens. After few mins/hours, it automatically resolves.
We have contacted MS support and could not get anything to get this resolved
Has Anybody faced a similar issue ?
This does not constitute an answer but it won't fit in a comment!
Does this affect more than one report?
If more than one, do they all fail at the same time or will one work and another not when you try to run the at the same time?
If the situation is "all fail" or "all work" at any single point in time then try adding a new report with no parameters.
If this attempts to run when other reports fail then it sounds like the datasource cannot be resolved or a timeout connecting to the database.
If your parameters are populated by querying the database the report will 'hang' until it can complete the queries to populate the parameters so by running a parameter-less report you should see an error instead.
Try running a trace on the database server to see if you can see the queries being run at all.
A bit vague I know but it might help you understand where the problem lies a little better.
Since recently I'm getting an error in a Dynamics 365 updated CRM OL when rendering a report. Now when I test in VS, it also shows this error. It says the following:
Could not create a list of fields for the query. verify that you can connect to the data source and that your query syntax is correct. The specified record type does not exist in Microsoft Dynamics CRM. The entity "XXX" cannot be found. Specify a valid query and try again. The entity with a name = 'XXX' was not found in the MetadataCache
I can connect, this always worked using the CRM OL Link, credentials are the same and the entity is there. I just copied the fetchxml from my advanced find, so pretty sure it's there.
I don't know of any changes to 365 that should break this. However, what I would do to fix it is create a report using the wizard that uses this same entity. Download that report and check the following:
Data Source - Are there any slight differences in Data Source name and/or connection string?
Data Set - Are there any subtle differences in the FetchXML from casing to versions etc?
That should point out the problem. If not, and that downloaded report works, either just copy in the attributes etc you need into the new report and go with that instead (it might be some xml within the report itself). Or if your report has nothing on it you could start comparing the raw xml of the 2 reports to see what is different.
This error: "The entity with a name = 'XXX' was not found in the MetadataCache" is quite common error with caching in CRM (Metadata is cached to speed things up). On On-Premise environemtns the thing that usually helps is simple IISRESET, which is not an option for CRM Online. I would try obvious solutions like re-publishing all customizations in CRM (from my experience it rarely helps for this error, but it's worth a shot), maybe adding some dummy plugin (and then removing it). If you will be still getting this error, then contacting MS Support should be your next move (also IIS AppPools are resetted after some time, so maybe tomorrow it will work - seems like a bad joke but actually it's true...)
If you have multiple instances of CRM make sure you point to the correct one but adding ;crminstance; at the end of the connection string. This solved the problem for me.
Morning folks.
This problem is very isolated but very annoying. It only happens with one customer I do work for, and can happen when creating new datasets or amending existing ones, but doesnt happen every time.
So lets start with a new dataset. I right-click and Add Dataset, I give the dataset a name, select the data source and then select Stored Procedure. This is where the fun begins. I start to type the SP name and BOOM, Report Builder crashes. I know I can just pick off the list but in the case of this client, the list of SPs is massive, so starting to type the name narrows down the list.
This can also happen if I amended the name of an SP within an existing dataset.
It has to be something to do with configuration as this doesnt happen with any of my other customers. The only difference with this site is they use Citrix but I cant see how that would affect this.
So, Report Builder 3.0 connecting to a SQL Server 2008 R2 instance. Any suggestions?
Thanks in advance
It seems not the issue on SSRS side, please let the user connect the database and execute the SP directly to check the performance. It also can be some hardware or configuration issue on that client's machine. You can suggest him use Process Monitor to fetch some detail information.
Good morning, All.
First let me start off by saying that I'm extremely new at CRM and even after reading umpteen million articles, whitepapers, and blog entries, I still feel completely lost.
I have an instance of Dynamics CRM 2013 On Premise that I'm trying to write custom reports for. Before I realized that all reports had to be done inside of BIDS, I wrote out a beautiful custom Quote inside of SSRS itself. I made sure to use the Filtered Views in my query to the database, and the structure of the query seems sound, but I can't seem to upload the .rdl file into CRM.
I get the error:
Reporting Error
Error occurred while setting the data source for the report
I have two questions:
How do I move this report into CRM without having to fully recreate it in BIDS?
How do I pass the Quote ID from CRM to this report query?
Thanks in advance for all of your help.
Edit: Added Error Message
You Should be able to upload the RDL directly to CRM...
Try creating a really basic report with the same data source and something along the lines of "select top 10 from filteredincident" and upload it.
If it works then you know it's something with your query.
CRM is notorious for giving error messages that don't have enough information, or are at worst misleading. I've seen this error when I accidentally used non filtered tables and a autofilter parameter that was incorrect.
Also take a look at the SSRS server that CRM talks to, if it's misconfigured that may cause an issue as well. You may want to try uploading your report to SSRS and run it from there to ensure that SSRS is working correctly.
I have created a SSAS project and it works perfectly fine my cubes, data-sources and such first time I deploy the cube, but if I change the data in my database and I click Cube>Process in BIDS it will not reflect the new changes even though when I retrieve the table data in SQL server Management Studio the table shows my data is changed.
I have also tried updating the cube in SSIS package using the Analysis Services Processing Task. However, the changes in my underlying data is not shown. It stays the same. Can anybody give me the few possible scenarios that can cause this problem.
Much appreciated
Thanks in Advance
First step is to verify that the datasource ON THE SERVER (not locally on your dev machine) is set to the correct database.
Are you processing the whole project or just the cube? I noticed that I have to process the top level item in the solution explorer. Processing just the cube was not enough.
Do you get any error messages?
Try processing the dimensions first and then process the cube.
I am not an expert myself, but I understood that there is a difference between updating ("process") and "deploy".
So when you create or add dimensions you will first have to deploy the cube to the analysis server and then process it.
Later when there is new data in your datasource, you can just process the cube (it's already deployed) to update the data.
Try this:
Identify which Dimension or Fact Table data you have updated.
Goto the SSAS project you have created in Visual Studio for the CUBE.
In the Solution Explorer expand Dimensions folder.
Right Click the Dimension or Dimension related fact table for which data is updated and Click on Process.
Then once the Process window is opened then do confirm in the window then Object Name = and Process Option = "Process Update".
Then Click on run on the bottom of the Process window.
Verify the process is succeed, if failed then fix the errors.
Go to Browser Tab of the CUBE and check the data is updated.
Any changes made to the cube structure i.e. adding new measure(s) or business intel. will reflect merely after processing the cube, there is no need to deploy the whole project again. However if any changes are made in underlying data or data structure, then the whole project needs to be deployed.
Changes can be seen right after processing/deployment in Cube>Browser but will take some time to be realized in SSRS because of caching.