Memory issues while fetching content

HI all,
I am using SAP KM Apis. however i am facing some memory issues while using them.
I am using getContent() api on a resource. and then getInputStream on the Icontent object. I am using a ByteArrayOutputStream for writing the data. here is the code snippet:
byte[] buf= new byte[4096];
InputStream inputStream = content.getInputStream();
ByteArrayOutputStream os = new ByteArrayOutputStream(1024)
while ((count = inputStream.read(buf)) != -1)
    os.write(buf, 0, count);
byte[] arr=os.toByteArray();
THis code consuimes a lot of memory . we need to optimize this. we also tried removing BYteArrayOutputSteam and directly read the bytes in chunk :
int n; 
do 
     n = inputStream.read(buf, 0,(int)content.getContentLength()); 
while (n != -1); 
however this approach has not helped us.
please suggest an approach where memory consumption is less and the entire content is fed.
thanks

Hi
This code worked for me for reading from a KM res...
BufferedReader reader = new BufferedReader(new InputStreamReader(is));
StringBuilder sb = new StringBuilder();
String line = null;
        try {
            while ((line = reader.readLine()) != null) {
                sb.append(line + "\n");
        } catch (IOException e) {
            e.printStackTrace();
        } finally {
            try {
                is.close();
            } catch (IOException e) {
                e.printStackTrace();
return sb.toString();
Regards
BP

Similar Messages

  • Issue while fetching the file through *.extension by FTP sender file adapte

    Hello Experts,
    I am facing a issue while fetching the data through sender File adapter with  ' *.file extension' .I am illustarting the scenario as below .
    It is a simple scenarion of File to File inboumd scenarion.Here file is getting picked up from a third party system
    through FTP sender chanel and stored in a temp folder of PI through NFS rceiver file adapter .
        The problem is however I am getting, while picking the file with file name as "*.exo"(where exo is the file extension).
    But while fetching the file with particular name like"abcd_10032011*.exo"(file naming has been done by combination of abcd(always same)_currentdate(change according to current date)),file picked successfully .
    So here ,in the prior case file not getting picked up,but in later case it dose .
    Can anyone please let me know what might be the issue?

    Hi Sunit,
    Connect from your PI System to 3rd Party System (where are placed the Source Files)
    FTP <PartySystemHostName>
    eg. FTP 10.2.3.456 (then insert Username & Password to Login)
    Go to source directory
    cd \<SourceDirectory>
    eg. cd \donaldduck\directory\
    Execute a File List command
    ls -la *.<extension>
    eg. ls -la *.exo
    In this way you should be able to view all files with this extension (*.exo), the same action that Sap XI perform to pickup the file.
    Then, try to copy that file to your Local PI System (to check if there're some permissions issue):
    mget <filename>.exo
    eg. mpget File1_01012011.exo

  • Hyperion IR : Getting out of memory error while fetching data for whole year through web client (wrokspace)

    Hi,
    While fetching data though IR wen client from workspace for a year(all 12 months) I am getting error as ("Out of Memory .Advice : Close other applications or windows and try again").
    If I am trying same through IR studio it does not give any output and show me same repoting front page.
    If i am selecting periods till 8 months it is giving the required data in both IR web client and IR studio.
    Could you please suggest how can we resolve this issue.
    Thanks,
    D.N.Rana

    Issue Cause :
    Sometimes this is due to excessive data which brings the size of the BQY file up around one gigabyte uncompressed in size (for processing may take twice as actual RAM, plus the memory space space for the plugin, and the typical memory limit on a 32-bit system is 2 gigabytes).
    Solution :
    To avoid excessive BQY size exceeding memory availability:
    Ensure that your computer has at least 2Gb of free RAM before he runs IR Studio.
    Put a limit to the number of rows that can be pulled down: Right click on Request label of Query section and put a value in Return First xxx Rows (and check the check box).
    Do not pull down more than 750 MB of data (remember it may be duplicated while processing).
    Place limits or aggregations in Query section (as opposed to Result section) to limit data entering the BQY.

  • Error While fetching content nodes from Weblogic Portal

    Hi,
    I am trying to deploy an application on Weblogic server which will connect to Weblogic portal repository and fetch content nodes from the specified repository. While accessing the repositroy nodes from my jsp code I get following error:
    I am not very clear what is going wrong
    Error 500--Internal Server Error
    com.bea.content.RepositoryRuntimeException: Repository configuration error.
         at com.bea.content.manager.internal.RepositoryManagerImpl.getRepositoryOps(RepositoryManagerImpl.java:194)
         at com.bea.content.manager.internal.RepositoryManagerImpl.connect(RepositoryManagerImpl.java:141)
         at com.bea.content.manager.RepositoryManagerFactory.connect(RepositoryManagerFactory.java:46)
         at com.bea.content.federated.internal.VirtualRepositoryManagerImpl.getRepositoryManager(VirtualRepositoryManagerImpl.java:306)
         at com.bea.content.federated.internal.VirtualRepositoryManagerImpl.getSearchOps(VirtualRepositoryManagerImpl.java:332)
         at com.bea.content.federated.internal.itemloader.NodeSearchItemLoader.fetchItemIdentifiers(NodeSearchItemLoader.java:68)
         at com.bea.content.federated.internal.itemloader.DirectAccessItemLoaderBase.minimalResolve(DirectAccessItemLoaderBase.java:17)
         at com.bea.content.paging.internal.PagedListImpl.(PagedListImpl.java:89)
         at com.bea.content.federated.internal.SearchManagerImpl.search(SearchManagerImpl.java:55)
         at jsp_servlet.__test._jspService(__test.java:133)
         at weblogic.servlet.jsp.JspBase.service(JspBase.java:34)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:226)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:124)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:283)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3370)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(Lweblogic.security.acl.internal.AuthenticatedSubject;Lweblogic.security.acl.internal.AuthenticatedSubject;Ljava.security.PrivilegedAction;)Ljava.lang.Object;(Unknown Source)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:2117)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:2023)
    Caused by: javax.naming.NameNotFoundException: While trying to lookup 'examplesWebApp.BEA_content.RepositoryOpsHome' didn't find subcontext 'examplesWebApp'. Resolved ''; remaining name 'examplesWebApp/BEA_content/RepositoryOpsHome'
         at weblogic.jndi.internal.BasicNamingNode.newNameNotFoundException(BasicNamingNode.java:1138)
         at weblogic.jndi.internal.BasicNamingNode.lookupHere(BasicNamingNode.java:246)
         at weblogic.jndi.internal.ServerNamingNode.lookupHere(ServerNamingNode.java:171)
         at weblogic.jndi.internal.BasicNamingNode.lookup(BasicNamingNode.java:205)
         at weblogic.jndi.internal.WLEventContextImpl.lookup(WLEventContextImpl.java:254)
         at weblogic.jndi.internal.WLContextImpl.lookup(WLContextImpl.java:367)
         at javax.naming.InitialContext.lookup(InitialContext.java:351)
         at com.bea.content.ContentEJBHomeCache.lookup(ContentEJBHomeCache.java:56)
         at com.bea.content.manager.internal.RepositoryManagerImpl.getRepositoryOps(RepositoryManagerImpl.java:187)
         at com.bea.content.manager.internal.RepositoryManagerImpl.connect(RepositoryManagerImpl.java:141)
         at com.bea.content.manager.RepositoryManagerFactory.connect(RepositoryManagerFactory.java:46)
         at com.bea.content.federated.internal.VirtualRepositoryManagerImpl.getRepositoryManager(VirtualRepositoryManagerImpl.java:306)
         at com.bea.content.federated.internal.VirtualRepositoryManagerImpl.getSearchOps(VirtualRepositoryManagerImpl.java:332)
         at com.bea.content.federated.internal.itemloader.NodeSearchItemLoader.fetchItemIdentifiers(NodeSearchItemLoader.java:68)
         at com.bea.content.federated.internal.itemloader.DirectAccessItemLoaderBase.minimalResolve(DirectAccessItemLoaderBase.java:17)
         at com.bea.content.paging.internal.PagedListImpl.(PagedListImpl.java:89)
         at com.bea.content.federated.internal.SearchManagerImpl.search(SearchManagerImpl.java:55)
         at jsp_servlet.__test._jspService(__test.java:133)
         at weblogic.servlet.jsp.JspBase.service(JspBase.java:34)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:226)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:124)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:283)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3370)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(Lweblogic.security.acl.internal.AuthenticatedSubject;Lweblogic.security.acl.internal.AuthenticatedSubject;Ljava.security.PrivilegedAction;)Ljava.lang.Object;(Unknown Source)

    Repositories are essentially external to applications.
    Applications are associated with repositories via repository configs. Repository configs allow applications to access repositories.
    1. There is no special 'centralized repository'.
    2. This can be done, but has caveats, especially around caching. If multiple applications have repo configs which refer to the same repository, there can be caching issues. For example, suppose apps A and B both have repo configs for accessing repo R. If you use app A to change some data in repo R, then try to read the change via app B, you can end up with cached data instead of the most recent data in repo R. This is something to be aware of with this type of configuration.
    3. There is typically one WLP admin url per application. For example, if you have an application named 'FooApp', then the admin url can typically be found at 'http://localhost:7001/FooAppAdmin'. Application 'BarApp' may have an admin url at http://localhost:7001/BarAppAdmin.
    You can use the weblogic server console at http://localhost:7001/console to see these -- click on Deployments, then click Monitoriing. In the Context Root column, you can see the various top-level context roots registered for this application.
    4. Repositories are top-level possibly-shared entities, similar to databases.
    Applications 'use' Repositories via Repository Configs.
    Applications 'own' RepositoryConfigs, which allow them to 'use' a specific Repository.
    It's possible to have multiple applications 'use' the same repository (each application would have to have its own RepoConfig referring to the same repository), but there can be caching issues with this configuration -- see (2).
    Each application 'owns' its own set of repo configs, and therefore its own VCR namespace. The VCR namespace is compsed of the RepoConfigs that application 'owns'. For example, if an application has repository configs with names 'Repo1' and 'Repo2', then the VCR namespace includes /Repo1/... and /Repo2/...
    Note that each RepoConfig entry has a name, which becomes its VCR namespace path. The path to a node in a repository begins with / followed by the name of the RepoConfig followed by /. Application A and application B may each be accessing Repo R, but via different names. For example, if A has the path to Repo R registered via name 'RepoR', then its VCR path to Repo R is /RepoR/... If B has the path to Repo R registered via the name 'R', then its VCR path to Repo R is /R/...

  • Not enough memory issue while acquiring data

    Hi all,
    I am trying to acquire audio data @ 50KS/s for 20 sec. using the PXI 4498 card in the PXI system. To acquire data I am using the DAQmx functions.
    Step1 : Created DAQmx channel
    Step2: Set the no. of sample to read as 20X50=1000000 to the DAQmx Timing function
    Step3: DAQmx Read funtion to read the data. (Set the timeout as 20x2=40Seconds)
    PLease find the attached screen shot of the block diagram.
    Issue is sometimes I am facing the memory issue when I run this VI. Popup saying that "Not enough memory to complete this Operation"  is getting popup and the Vi getting stop abruptly. Kinldy look into the screen shot and suggest me the apprapriate method to read the data at mentioned acquisition rate.
    Note: The problem is not consistent. Sometimes I am facing the error and sometimes not.
    Thanks in advance
    Giridhar
    Attachments:
    Block diagram.jpg ‏208 KB

    I think actdi is talking about loading an external database directly from the XML file using the jdbc XML driver, correct? I have not tried this yet, but am curious as to the results.
    http://docs.oracle.com/cd/E17904_01/integrate.1111/e12644/appendix_xml_driver.htm#CHDFIJEH
    Hope this works.
    Enjoy!
    Michael R.

  • LabVIEW Memory full while fetching data from database

    Hi,
    In my program I need to sync some data from client PC as per the selected time frame.
    But while fecthing the data from the clinet database, my application is hanging and when I run it with code I get the 'LabVIEW memory full' error message.
    Kindly suggest to overcome this problem.

    Fetching the entire database is probably not a good idea.  You should narrow down how much you read at a time.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Issues while watching content after live streaming operation

    Hi,
    1)We have performed a live streaming of an event using Azure Media Services , we streamed the entire event 
    which was of about 5-6 hrs duration . But at the end we came to know that it has streamed only last  1 hour of 
    entire event.
    Can you let me know what might be the issue?
    2)We performed a live streaming of an event in 2 parts using azure media services. After 2 hours we stopped the streaming of 1st part and deleted the channel for 1st part and we started streaming of another part by creating another channel . We are unable to
    view the 1st part through content url while we are able to view the 2nd part through content url?
    Can you let me know what might be the issue?
    Thanks,
    Pushkar

    The fan is speeding up to cool the GPU. The GPU (graphic's chip) heats up because video's and game's work the GPU much harder than normal desktop activities.
    When was the last time you checked and cleaned the air intake grill on the bottom of your iMac?
    I shutdown (stopping the fans), disconnect the power and vigorously vacuum out the bottom grill work on my Mac's a couple of times a year to keep them running cool.

  • Memory issues while running a yearly report

    Hi,
    One of my consultant is trying to run an yearly report which consumes approximately 3 GB of size. But, the memory available at that point is only 2 GB on the server. The report generates a runtime error "SYSTEM_NO_TASK_STORAGE".
    But, when I restart the server, all the buffers are cleared and there is 5 GB space after which the report runs to success. The number of program swaps in ST02 are more than 3000. The swap space assigned is 400 MB.
    I believe that increasing the swap space can resolve this issue. But, I would like to know any other better solutions for this.
    Thanks
    Vijay

    Hi,
    Of course, the best solution would be to add some physical RAM on your server.
    Nowadays, I would say that using less than 8 GB RAM for SAP products leeds running into problems.
    In my company, we are now going full 64bit with at least 16 GB RAM...
    Bythe way 400 MB for the program buffer seems very low to me. On SAP R/3 4.7 we use 1500 MB. But this just slows down the system.
    You should check your extended memory setting.
    Regards,
    Olivier

  • Out Of Memory Issue while dowloading infomraiton from DB to XML

    Hi,
    I am converting the Database tables into XML file format
    using Java IO and SAXP. I am keeping all XML files in
    download folder. But, while downloading process
    i am getting Out Of Memory Error.
    When i tried to download the real data, the download
    folder size is going tp 50 MB and i am getting Out Of
    memory error?
    Is that related to JVM Memory or System Memory. What
    would be the solution for this?
    Awaiting your answers

    By default the JVM appropriates 96Mb of memory for it's heap. You can increase the allocation by putting, say -Xmx128m on the Java command line, though perhaps a better solution is to avoid loading all your data at one time and write the XML out while reading the database.

  • 7510 Supernova - Memory Issues while Online!?

    I just got this Nokia 7510 Supernova and have noticed from the get go that every time I access the internet, while ready to flip to page two or three, OUT OF MEMORY appears. Sometimes a few navi oks will send it away other times, the entire phone crashes and reboots. Anyone else have this problem? I have full WEBtoGO service and there are no pictures taken or songs/tunes downloaded yet as it's brand new. The phone also had to Movie Trailers on it - very nice but I moved them to my sim card and now they won't play because "THERE IS NO VALID LICENSE" What's up with that? I tried to move them back to phone memory but was unable to because there wasn't enough space!!!???
    still livin in the 80's

    I just got this Nokia 7510 Supernova and have noticed from the get go that every time I access the internet, while ready to flip to page two or three, OUT OF MEMORY appears. Sometimes a few navi oks will send it away other times, the entire phone crashes and reboots. Anyone else have this problem? I have full WEBtoGO service and there are no pictures taken or songs/tunes downloaded yet as it's brand new. The phone also had to Movie Trailers on it - very nice but I moved them to my sim card and now they won't play because "THERE IS NO VALID LICENSE" What's up with that? I tried to move them back to phone memory but was unable to because there wasn't enough space!!!???
    still livin in the 80's

  • Performance issue while fetching metadata from Informatica Repository

    I'm working on Informatica 8.6(ETL tool) which contains its own repository to save metadata and using their Mapiing SDK API's I'm developing a java application which fetches the objects from the repository.
    For this purpose by using "mapfwk.jar", i'm first connecting to the repository using RepositoryConnectionManager class and then at the time of fetching the metadata i used getFolder, getSource & getTarget functions.
    Issue: Program is taking to much time for fetching the folders. The time taken by it depends on the number of metadata objects present in it,i.e. as object number increases, time increases.
    Please advise how to reduce time for fetching metadata from repository.
    Source Code:
    #1 - Code for connecting to repository
    protected static String PC_CLIENT_INSTALL_PATH = "E:\\Informatica\\PowerCenter8.6.0\\client\\bin";
         protected static String TARGET_REPO_NAME = "test_rep";
         protected static String REPO_SERVER_HOST = "blrdxp-nfadate";
         protected static String REPO_SERVER_PORT = "6001";
         protected static String ADMIN_USERNAME = "Administrator";
         protected static String ADMIN_PASSWORD = "Administrator";
         protected static String REPO_SERVER_DOMAIN_NAME = "Domain_blrdxp-nfadate";
    protected void initializeRepositoryProps(){
    CachedRepositoryConnectionManager rpMgr = new CachedRepositoryConnectionManager(new PmrepRepositoryConnectionManager());
                      RepoProperties repoProp = new RepoProperties();
                      repoProp.setProperty(RepoPropsConstant.PC_CLIENT_INSTALL_PATH, PC_CLIENT_INSTALL_PATH);
                      repoProp.setProperty(RepoPropsConstant.TARGET_REPO_NAME, TARGET_REPO_NAME);
                      repoProp.setProperty(RepoPropsConstant.REPO_SERVER_DOMAIN_NAME, REPO_SERVER_DOMAIN_NAME);
                      repoProp.setProperty(RepoPropsConstant.REPO_SERVER_HOST, REPO_SERVER_HOST);
                      repoProp.setProperty(RepoPropsConstant.REPO_SERVER_PORT, REPO_SERVER_PORT);
                      repoProp.setProperty(RepoPropsConstant.ADMIN_USERNAME, ADMIN_USERNAME);
                      repoProp.setProperty(RepoPropsConstant.ADMIN_PASSWORD, ADMIN_PASSWORD);
                      rep.setProperties(repoProp);
                      rep.setRepositoryConnectionManager(rpMgr);
    }#2 - Code for fetching metadata
    Vector<Folder> rep_FetchedFolders = new Vector<Folder>();
    public void fetchRepositoryFolders(){
              initializeRepositoryProps();
              System.out.println("Repository Properties set");
              //To fetch Folder
                    Vector<Folder> folders = new Vector<Folder>();
              folders = (Vector<Folder>)rep.getFolder();
                    for(int i=1 ; i < folders.size(); i++){
               Folder t_folder = new Folder(); 
               t_folder.setName(((Folder)folders.get(i)).getName());
               Vector listOfSources = ((Folder)folders.get((i))).getSource();
                            //To fetch Sources from folder
                   for(int b=0; b<listOfSources.size();b++){
                    Source src = ((Source)listOfSources.get(b));
                    t_folder.addSource(src);
              Vector listOfTargets = ((Folder)folders.get((i))).getTarget();
                            //To fetch Sources from folder
                   for(int b=0; b<listOfTargets.size();b++){
                     Target trg = ((Target)listOfTargets.get(b));
                     t_folder.addTarget(trg);
                   rep_FetchedFolders.addElement(t_folder);
              }

    Hi neel007,
    Just use a List instead of a Vector, it's more performant :
    List<Folder> rep_FetchedFolders = new ArrayList<Folder>();If you need to synchronize your list, then
    List<Folder> rep_FetchedFolders = Collections.synchronizedList(new ArrayList<Folder>());Also, if you're using Java 5 or higher and if you're sure listOfTargets contains only Target objects, instead of this
    for(int b=0; b<listOfTargets.size();b++){
       Target trg = ((Target)listOfTargets.get(b));
       t_folder.addTarget(trg);
    }you may do this :
    for (Target trg : listOfTargets) {
       t_folder.addTarget(trg);
    }Edited by: Chicon on May 18, 2009 7:29 AM
    Edited by: Chicon on May 18, 2009 7:35 AM

  • Regarding Memory issues((error) while scheduling the job

    Hi Friends
    I am  facing memory issue while rerunning BODS jobs in production.As we have rapid mart when jobs got failed i rerun the
    job suddenly job got failed and  in logs i found Memory issue
    please help me what i have to do in steps
    Thanks

    I think there is no one solution to buffer pool issue. Buffer pool issue happens due to many reasons like how you design the data flow, requirement of push down or even insufficient memory in the running environment.
    You can check by checking 'collect statistics for optimization' during running the job. Also in the data flow, try changing from pageable cache to memory cache by right clicking and selecting properties.  The below link give some details on caching, might be helpful.
    https://wiki.sdn.sap.com/wiki/display/BOBJ/CachinginDI
    Arun

  • Memory dump while loading data

    Hello,
    We are facing memory issues while loading data in BW system.
    This particular loading process is taking more than 4GB (3GB of EM and 1GB of heap) and still failing.
    We cannot increase the memory beyond this. Can you please suggest a way to load the data ?
    Runtime Errors         TSV_TNEW_PAGE_ALLOC_FAILED
    Date and Time          09.03.2010 22:15:49
    Error analysis
         The internal table "{A:2*\TYPE=%_T00003S00000114O0000015636}" could not be
          further extended. To enable
         error handling, the table had to be delete before this log was written.
         As a result, the table is displayed further down or, if you branch to
         the ABAP Debugger, with 0 rows.
         At the time of the termination, the following data was determined for
         the relevant internal table:
         Memory location: "Session memory"
         Row width: 1140
         Number of rows: 30824
         Allocated rows: 30824
         Newly requested rows: 8 (in 1 blocks)
    How to correct the error
         The amount of storage space (in bytes) filled at termination time was:
         Roll area...................... 14627872
         Extended memory (EM)........... 3004121016
         Assigned memory (HEAP)......... 1000024976
         Short area..................... " "
         Paging area.................... 24576
         Maximum address space.......... 4294967295
         If the error occures in a non-modified SAP program, you may be able to
         find an interim solution in an SAP Note.
         If you have access to SAP Notes, carry out a search with the following
         keywords:
         "TSV_TNEW_PAGE_ALLOC_FAILED" " "
         "GP44XYX8R5ZMOGVGGMHG0U1U04O" or "GP44XYX8R5ZMOGVGGMHG0U1U04O"
         "CONVERT_FROM_MEMORY"

    Try to load by reducing data package size.
    I think you have some routine in transformation/update rule. Check that whether you are using "SELECT *" in that. It may be one reason for this error.
    You can check OSS Note 396642, 379266 and 997930.
    Edited by: Pravender on Mar 10, 2010 1:09 PM

  • Issue while playing the content of Azure Media Services after deleting the channel.

    Hi,
    We are having the issue while playing the content of Azure Media Services after deleting the channel.
    Please find detail scenarios below:
    We have done live streaming of an event using Azure Media Services, event was around 5-6 hrs. After completion of event channel was deleted, when we checked content data
    it was only 1hr. Content has only last 1 hr. recording, we are unable to get full recoding.
    We divided event into two parts to avoid above issue. After completion of 1st part which was of 2 hrs., we stopped the streaming and deleted the channel. New channel was
    created and streaming was started for 2<sup>nd</sup> part, after completion of event channel was deleted. Now we are unable to play the content data for 1st part but able to play the 2nd part of event through content publish URL.
    We have done streaming of 3 hrs. event on 18th April with same process and full recoding was available. 
    Can you let me know what might be the issue?

    Hi,
    Please refer to Justin’s blog and see if it helps:
    http://www.justinkobel.com/post/2012/04/30/SharePoint-2010-Claims%E2%80%93Operation-is-not-valid-due-to-the-current-state-of-the-object.aspx
    As suggested by Alex, please provide more information about your process of changing password.
    Here is an article for updating passwords on SharePoint 2010:
    http://blogs.technet.com/b/seanearp/archive/2011/01/25/updating-passwords-on-sharepoint-2010.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support

  • Authorization issue while using business content objects

    Hi all,
    I am getting an authorization error while loading from DSO to Cube (for standard business content objects only). Even i am not able access the data in Bex reports from Standard cubes or DSO.
    The user is assigned to SAP_ALL .
    But there is no such issue while accessing user defined DSO or cubes.
    any solutions will be helpful.
    Regards,
    Varma

    Hi,
    Have you seen the error thru the SU53 t-code ? There you should see what is the missing authorization with the user.
    Regards, Federico

Maybe you are looking for

  • Agent Determination Rule Not Working

    I have created an new agent determination rule in PFAC. The category I used is "Agent Determination: Responsibilities". I have also selected the "Terminate if Rule Resolution has no Result" option. The container has one element in it which is NOT man

  • Problems getting a higher framebuffer refresh rate

    Hello everyone just changed from Gentoo to Arch    Anyway here is my problem, I usually use a framebuffer resolution of 1280x1024-32@75. I used to use a Nvidia BFG Geforce Ti4200SE 128MB and it had no problem of booting into the above resolution. I r

  • BOP Monitor missing Sales Orders, Displays only STOs

    Hi, We have an issue with a back order processing monitor screen.  BOP Monitor displays both SO and STO immediately after the BOP run but wipes out the SO from display after few hours. I guess the SOs have been deleted from buffer and so we don't see

  • How to use smileys in message in N79, please help

    i had other nokia models wher in i used to send smileys in messages, but now i have taken N79 and there is no smileys insertion possible in it. please tell me how can i do so. if any software download is required then also please tell me

  • IPad and Wi-Fi problem

    Hi My iPad is showing a connection with my computer and my computer is also showing a connection withmy iPad but Safari asys can't open page because it is not connected to the internet. And in general on the iPad it shows Network Not available of cor