We have a reporting system where the default rendering format is HTML.
HOwever, in some cases user may export the data into Excel aftergenerating
the report in HTML. Howveer, this export is taking too much time. eg 5500
row report in HTML takes around 8 minutes to export into excel. Is there a
workaround for this? Please note that default rendering has to be in HTML
only.
Also another feature noticed is that in RS 2005, the report server execution
log seems to be logging seperate entries for export feature as well. This
was not happening in RS2000. Is this a new feature in RS2005 or is the
underlying SP for the new report being called again when the export to excel
happens?
i have a report with date formats,numbers, percentages,when runnig the preview it successfully those are visible in preview, but problem is when iam importing report into excel , those date formats,numbers,percentages are not visible in excel sheet. instead of date formats, numbers, percentages, cells are empty and cells not contain any data like dateformat , number,percentage.
so please help me because i have only one day for delivering repport,
i hope any body save me this problem, iam giving advanced thanks to our forum members.
Hi, My application needs to retrieve data from a table which has more than 15 lakh records. The records keep increasing in thousands every 15 days. Is there anyway i can reduce the time to retrieve? basically i have a select statement with a few conditions and a clause for the id's of these records.
i have a client who is dissatisfied with the merged columns and rows produced when SSRS renders to Excel, it prevents them from easily copying and pasting data into another workbook.
Is there going to be any work done in this area of the product in the near future? My understanding is that this 'issue' has been around for quite some time, and shortly after the SSRS2005 RTM release the team were going to look at tidying up the Excel rendering. Can anyone from MS make an official comment on this?
Ok guys, here's the problem. Normally if I were sorting a report that needed to be rendered to pdf or excel, I'll sort the data in the dataset before loading it into the reportviewer in my code. This is in vb, for example:
But for this particular case, I have to load the data into the report first, because in the actual report I have to combine two columns based on certain criteria (i do this using an expression for a column in my table) , and after those two columns are combined then they must be sorted.
But since the data is already loaded into the reportviewer I can't sort like I normally would, the report is rendered in either excel or pdf so interactive sort is pointless. I can't sort through table or properties or the properties of the text box because both fields are combined and you can only sort based on one field.
i have developed reports which uses Line textbox, and list ...
When i render these reports in acrobat and excel, acrobat works fine but excel it takes line every thicker even if i decrease the width of the line to 0.25pt excel takes as same as if it was 1.0pt
is there any solution for this, my report has no warnings .
my report dont fit on a A4 letter. thats not a problem at the report viewer but when i print or export the report. can i scale the report to fit to one page??
Okay, when will SRS 2005 support Office 2007 extensions.
I find it hard to believe that a service pack has not been releaseed to support Office 2007. If there is I apologize in advance for this message.
Info: Microsoft.ReportingServices.ReportProcessing.WrapperReportRenderingException: An error occurred during rendering of the report. ---> Microsoft.ReportingServices.ReportRendering.ReportRenderingException: An error occurred during rendering of the report. ---> System.Exception: Excel Rendering Extension : Number of rows in the Excel sheet exceeded the limit of 65536 rows.
When using a subscription to send mail, I find that the report can be sent in any format apart from Excel, which gives me the error: "Failure sending mail: Unspecified error". As you can imagine, this doesn't give me much to go on!
I want to use Excel format as it retains the drilldown functionality (hiding / unhiding rows). The ExecutionLog has the status as rsSuccess, but Subscription tab for the report shows the error.
I have a report which takes around 5 seconds when run in BIDS but takes around 20 seconds when deployed on report server.Execution log says TimeDataRetrieval is around 3-4 seconds and  rendering time is around 15-17 secs.From this report I am passing 8 parameters to a  drill through report and there are 36 text boxes where I have defined these parameters for drill through action.All these parameters are populated in main dataset.When I deployed the same report without any drill through action and parameter, it takes 5 seconds.So I am suspecting that because of drill through parameters report is taking more rendering time on server.I am using 2008R2 and IE11.
Is it expected behavior that due to so many parameters for drill through action, report will take more rendering time?If yes, then why is it not taking same time in BIDS?
First of all, this is not in reference to using SRS (SQL Reporting Services) to render a report and then use one of the extensions to render the complete report as a pdf, tiff, excel etc. We have an opportunity to render a list of claims and then embed the supporting docs for each of the claims within the report. We don't have an issue referencing jpg and gif images via URLs and then rendering them within the report after the grid information. We do have an issue rendering tiff images within the report.
You can insert an image object into SRS at design time and have it render and you can convert a complete report to a tiff image but I cannot find a way to be able to render a tiff image when running the report. All you get is the red "x".
My question is has anyone encountered the same issue and, if so, what did you do to resolve the issue?
Hi, Its a common issue with reporting services, exporting a report with huge data to excel takes long time to render. Im facing the same issue. Trying to export a SSRS 2005 report to Excel 2003 takes very long time. Problem 1: The time taken to generate the excel report is pretty long (about 10 minutes) as the report runs to hundreds of pages. (The excel has about 30,000 rows and is ~15 MB) Problem 2: Once I open the excel and close it the size reduces to half of it. This is understood because of lot of characters values in the report that can be represented as single byte string, that€™s probably where Excel is making a difference. In Reporting Services it always write strings as 2-byte Unicode, but Excel will always try to compress to single byte when possible.
Am majorly concerned about the Problem 1. Any solutions would be highly appreciated. Thanks in advance.
Like the title says.. I am trying to import an Excel 2000 spreadsheet to SQL Server to do some manipulations/aggregations. The Excel file has several cells with long text (detailed product descriptions) in them. Since this is a quick Import/Export I am using the Wizard rather than the SSIS Designer. I specify the value of the field as being varchar(max), and this is confirmed by the Edit SQL button. However, every time I try to run it I get an error on the Details column, stating that "Text was truncated". I have tried changing it from varchar(max) to text (and also tried nvarchar(max) and ntext) but still get the same error. What the heck am I doing wrong? I cannot just get rid of the column in question because we need its contents. Is this some issue with Excel and not SQL Server? If so, is there any way to fix it?
I ran a matrix report and chose to export it to excel. It was taking a very long time and then the error message below popped up. I can export other matrix reports to excel. Can someone shed some light on where I might determine is the cause of the problem?
Thread was being aborted. Description: An unhandled exception occurred during the execution of the current web request. Please review the stack trace for more information about the error and where it originated in the code.
Exception Details: System.Exception: Thread was being aborted.
Source Error:
An unhandled exception was generated during the execution of the current web request. Information regarding the origin and location of the exception can be identified using the exception stack trace below. Stack Trace:
[Exception: Thread was being aborted.]
[Exception: An error occurred during rendering of the report.]
[Exception: An error occurred during rendering of the report.]
[Exception: The operation has timed out. The report server has canceled the operation.] Microsoft.Reporting.WebForms.ServerReport.ServerUrlRequest(Boolean isAbortable, String url, Stream outputStream, String& mimeType, String& fileNameExtension) +882 Microsoft.Reporting.WebForms.ServerReport.InternalRender(Boolean isAbortable, String format, String deviceInfo, NameValueCollection urlAccessParameters, Stream reportStream, String& mimeType, String& fileNameExtension) +1312 Microsoft.Reporting.WebForms.ServerReportControlSource.RenderReport(String format, String deviceInfo, NameValueCollection additionalParams, String& mimeType, String& fileExtension) +215 Microsoft.Reporting.WebForms.ExportOperation.PerformOperation(NameValueCollection urlQuery, HttpResponse response) +204 Microsoft.Reporting.WebForms.HttpHandler.ProcessRequest(HttpContext context) +137 System.Web.CallHandlerExecutionStep.System.Web.HttpApplication.IExecutionStep.Execute() +499 System.Web.HttpApplication.ExecuteStep(IExecutionStep step, Boolean& completedSynchronously) +177
I need help to fix SSIS Send Mail Task Error. I have a
Data Flow Read from Sql Data base and then put the query result into excel file (OLE DB Source --> Excel Destination)
Send Mail Task Email excel file from the data flow
Control Flow : Data Flow --> Send Mail Task
The package throws an error below. I tried to disable my antivirus because i thought it's throwing an error because of the antivirus. But, it's still throwing the same error. If i tried to send email without attachment, it works just fine. Is there anybody know how to fix this?
Progress: The SendMail task is initiated. - 0 percent complete [Send Mail Task] Error: An error occurred with the following error message: "The operation has timed out.". Progress: The SendMail task is completed. - 100 percent complete Task Send Mail Task failed
Hi,I have the following three tables below containing Resources,Categories and a link table so each Resource can belong to one or moreCategories. I would like to create a view (ResourceID, ResourceName,CategoryID, CategoryName) that includes one row for each Resource withjust one of the Categories that it belongs to.Resource table- ResourceID- ResourceName- etc..Category table- CategoryID- CategoryName- etc..ResourceCategory table- ResourceID- CategoryIDCan anyone help? Thanks.
i used "dbcc shrink file" to reduce the log file of a database.the query analyzer says "successfully executed" but the log file doesn't seem to reduce..am i missing something?
The transaction log of the databse i am using has grown up to 7GB...(previously the setting was unrestricted file growth...now changed to restrict file growth to 7 GB approx.) now this 7GB space is not needed....i would like to reduce the size to around 2 GB...how can i achieve this?i observer that on the properties i can only increase the size and not decrease it...also i am using transactional replication..this server is the publisher to four subscribers..
what i want to do, is to 1.- insert a row in User, 2.- insert a row in Car 3.- using SCOPE_IDENTITY, to insert a new row in user_car that relates the ids of the last added car with the last added user.
But i need that to be done in one single transaction. Once that is done, i have another question
how can i do the same but where i can add a variable number of different cars? in other words, i add 1 user and then add 5 cars, then create 5 rows in the user_car where i have the id of the last added user with the incremental id of the last 5 added cars
how can this be done in one single transaction, i cant make it in 3 transactions cuz it would cause me a lot of trouble. Any help?
my transaction should be something like this
Code:
insert into user(username) values('user1'); insert into car(carName) values('car1'); insert into user_car(???,???)
but im not sure how to do this in one single transaction
Hi, If I understand it correctly, you only need an LDF file to restore to a point in time after the last full backup? If this is so, then Could the LDF file not be reduced in size on perfoming a full backup? Most of the time it's not an issue as there is enough space on the HDD, but is it possible to reduce the ldf file size periodically (manually would be fine). Is changing the recovery mode from FULL to SIMPLE and then back to FULL an option? If so, is anyone able to tell me how, exactly, I can do this? ... I've sifted through the documentaion to no avail ... :eek: Many thanks Rob
hi i have a database that becomes to big after a few days. is there a chance to say if the table reaches a number of lines or a special memorysize, delete (or better archive) the oldest entries?
I have started to look at the way our production DB has been defined and set up, with the view to improving performance.
The DB is now 11gb, and the original size was set up as 3000mb, the rest has been take in 10% additional extents.
Now, back in my DB2 DBA days, this was a bad thing to have any data spread across extents as they may not be contiguous. I am assuming that is the same with SQL Server. Can someone confirm/deny this?
If this is the case, how can I get the DB back into one primary partition?
If I have a transaction log in a database of size 1GB ( space allocated is during creation of database) currently only 300 mb of its space is used i.e. nearly 700 mb is free. If I want to reduce physical file size of transaction log by 200 mb and release it for operating system then How can I do it???
I have inherited a number of databases which were substantially over sized when they were set up. I'd like to reduce both the log and database files to be smaller than their original sizes, what's the easiest way to do this? If anyone has any experience of doing this please reply.
i am new to sql server. i recently found the transaction log size of my database has reached 109 MB. how can i reduce it. a transaction log backup was sceduled daily at 12.00 noon nad full backup monthly.
I have a production database of a size of 70 GB. Half of the data was archived and deleted from the current database. What is the best way to reduce the size of the database, as we cannot shrink an entire database to be smaller than its original size? Thanks a lot!
I created few jobs that would archive the production DB and delete the archived data... but it looks like the DB size is not reducing!!! Some times it looks like the size has increased!!
I think this is because of the log file size has increaded by the DELETE operations....But what can I do for this???
Code: <root> <timesheet empid="1" entryyear="2012" entryPeriod="1" adminnotes="These are the admin notes" empnotes="These are the user notes"> <project projnum="TestProject" projname="The really big project for our best customer"> <activity actcode="000103020200302302322" actname="Demolish the 55th story of the main tower">
[Code] ....
Notice how there is a tremendous amount of redundancy in the XML. I was hoping to come up with an XML result of the following, which transmits the same data, without the redundancies.
Code: <root> <timesheet empid="1" entryyear="2012" entryPeriod="1" adminnotes="These are the admin notes" empnotes="These are the user notes"> <project projnum="TestProject" projname="The really big project for our best customer"> <activity actcode="000103020200302302322" actname="Demolish the 55th story of the main tower"> <expenditure expcode="1" expname="Regular Hours">
I'm trying to insert all the rows from a table to a new table. (insert A select * from AA) The reads on Profiler shows ar really high value (10253548).
First I created a unique clustered index and the reads shows (3258445), then I created a non clustered index expecting to have lower reads. Instead the reads shows (10253548).
I read creating indexes helps reduce reads. But it's not happening. Any ideas what is going on?