File Roll Behavior for Blob Storage Output

What is the file roll behavior when using blob storage output for a stream analytics job. That is, when is the current blob closed, and a new blob opened? I notice that when using json output, the blob is a json array of events. Is the blob valid json
while it is being written to?

Today, the file is rolled when it reaches a certain size, but it is not safe to rely on that behavior.   Do you have a need to control the roll behavior?
The blob is an almost valid JSON while it is being written to --- namely it is missing the closing ']'.  If you add it, it will parse.

Similar Messages

  • How to set CORS properties for BLOB Storage using node?

    Hi - I just got started with Azure using a Node-based web site and mobile services.
    I am following various documentation in order to provide an API for users to upload images via a time-restricted SAS for the BLOB Storage.
    In order to upload my image, I need to set the CORS configuration for the BLOB Storage. Unfortunately this cannot be done via the management portal.
    I'm unclear as to how to accomplish this. I'm considering using the startup.js file in my mobile service to make a post request to the BLOB Storage REST API:
    http://msdn.microsoft.com/en-us/library/windowsazure/hh452235.aspx
    Are there appropriate methods in the Node SDK to make this easier, especially the signing part?
    What is the recommended way for setting CORS properties for the BLOB Storage via Node?
    Thanks for your help
    Stefan

    Unfortunately Node SDK does not support CORS functionality yet. Your option would be to write code which consumes the REST API for setting CORS. Not sure if it helps but there's a free tool out there written by my company which you can use to set CORS
    on your storage account. More information about this tool can be found here:
    http://blog.cynapta.com/2013/12/cynapta-azure-cors-helper-free-tool-to-manage-cors-rules-for-windows-azure-blob-storage/
    Hope this helps.

  • Uploading Files to DMP for local storage (Failover ) in DMM/DMP 5.2 is not working

    Dear All,
    I face a problem in making failover on the DMP , I did the following:
    1- make the design and save the presentation the needed to be failover.
    2- from Advanced Tasks ---> File transfer to DMP or Server and select the presentation.
    3- in the Adanced Task ---> Go to URL you will find an automated file genrated .
    4- Go to Digtal Media Player and select the generated URL and select the DMP and press GO.
    in DMM 5.1 after the above steps,the player was restart and after it back it displayed the uploaded presentation.
    but now the system running with 5.2 and it is not working ,Can any one help me to make this in DMM/DMP 5.2?
    Thanks,
    Ahmed Ellboudy.

    Ahmed,
    Let's not worry about failover at this time.
    Let's just attempt to get the LOCAL presentation to play.
    Your steps appear to be correct for transferring the
    presentation to the DMP.
    * From the DMM-DSM, Select the action to play the presentation
      on the DMP in Question
    Does the Presentation play on the DMP?
    then
    * go to the advanced tasks on the DMM-DSM and find the
      GO TO URL for the presentation that you created.
      Copy URL to your clipboard.
    * Now go to the DMP-DM of the DMP where you transferred the
      presentation.   Go to the Display Actions-->URL to be displayed
      and Paste the GO TO URL here then press the GO button
    Does the Presentation play on the DMP?
    Let me know what you see....
    T.

  • File Adapter problem for Fixed Length output file

    Hi There!,
    The following is my sample data format
    ABCD 2345 03Accounting Swissland
    ABCD 011234 1000Jack Gates 400000 03
    ABCD 021234 1001Bill Welch 400000 03
    In the first record has dept details:
    DummyText, Row identifier(not given in case of dept), Identifier, Deptno, Deptname, Location
    The second and third records are emp details:
    DummyText,Row identifier, Identifier,EmptNo,Emp Name, Manager, Salary, Deptno.
    Notice that the primary key is defined by the first three columns DummyText, Row identifier, Identifier.
    When I use Native Format Builder from File Adapter and scan the first three columns I get all three records (i.e dept and both employees). I deleted the second employee and proceeded to finish the configuration. But, the problem I face is that the key value shows as "ABCD% 011234" for both the employee records where as I would like it to show as "ABCD 011234" for the first record and "ABCD 021234" for the second record.
    How do I get around this problem?
    Moreover if I have a group or recurring records and not just one record recurring. I mean assuming I have the following:
    First row: Department
    Second row: Employee
    Thrid row: Manager
    Fourth row: Employee
    Fifth row: Manager
    How do I get this done?
    Does anyone have an idea where to find details of nxsd.
    Thanks in Advance.

    Hi Manoj
    Give the FCC parameter as
    RecordSet Structure = Header,,Detaillnes,,Detaillines2,*
    Header.endSeparator = 'nl'
    Header.fieldFixedLengths = field length
    Header.fieldFixedLengthType = char/byte
    Header.fieldNames = field1,field2,field3,.....
    Header.keyFieldValue = Header
    Detaillnes.endSeparator = 'nl'
    Detaillnes.fieldFixedLengths = field length
    Detaillnes.fieldFixedLengthType = char/byte
    Detaillnes.fieldNames = field1,field2,field3,.....
    Detaillnes.keyFieldValue = Detaillnes
    Detaillines2.endSeparator = 'nl'
    Detaillines2.fieldFixedLengths = field length
    Detaillines2.fieldFixedLengthType = char/byte
    Detaillines2.fieldNames = field1,field2,field3,.....
    Detaillines2.keyFieldValue = Detaillines2
    The keyFieldValue should be the Value which identifies your record either as Header/Detaillines/Detaillnes2.
    Regards
    Santhosh
    Message was edited by:
            Santhosh Kumar V

  • File Content conversion for Fiexed length output file

    Hi,
    I am Working on Idoc to File scenario.
    I would like to know the content conversion partmenrts for the follwoing structure.
    Structrure is as below:
    Message Type
      Header Record  1..unbounded
    .    Field1   0...1
    .    Field2   0...1
    .    Field3   0...1
    ..    Detaillines   0...unbounded
    ...       Field3  0..1
    ...       Field4  0..1
    ..    Detaillines2   0..unbounded
    ...       Field6   0...1
    ...       Field7   0...1
    Can you please advise.
    Regards,
    Manoj Kumar

    Hi Manoj
    Give the FCC parameter as
    RecordSet Structure = Header,,Detaillnes,,Detaillines2,*
    Header.endSeparator = 'nl'
    Header.fieldFixedLengths = field length
    Header.fieldFixedLengthType = char/byte
    Header.fieldNames = field1,field2,field3,.....
    Header.keyFieldValue = Header
    Detaillnes.endSeparator = 'nl'
    Detaillnes.fieldFixedLengths = field length
    Detaillnes.fieldFixedLengthType = char/byte
    Detaillnes.fieldNames = field1,field2,field3,.....
    Detaillnes.keyFieldValue = Detaillnes
    Detaillines2.endSeparator = 'nl'
    Detaillines2.fieldFixedLengths = field length
    Detaillines2.fieldFixedLengthType = char/byte
    Detaillines2.fieldNames = field1,field2,field3,.....
    Detaillines2.keyFieldValue = Detaillines2
    The keyFieldValue should be the Value which identifies your record either as Header/Detaillines/Detaillnes2.
    Regards
    Santhosh
    Message was edited by:
            Santhosh Kumar V

  • Force file download for IE9 on Azure Blob Storage

    Hello,
    I am trying to use Azure Blob Storage as a location for secure file downloads using Shared Access Signature. Everything is working very well, however the problem I am having is I am trying to allow the user to save files from the browser and I have all browsers
    except IE9 working.
    Reviewing this question,
    What content type to force download of text response?
    this works well when I can control all of the headers, however in Azure Blob Storage, I have set the Content-Type to application/octet-stream and this allows all browsers except IE to ask the user to save the file, IE simply opens the file. It appears that
    known file types will open (example .jpg, .wmv etc…).
    In Azure, I have found no way to set
    Content-Disposition: attachment;filename="My Text File.txt"
    Is there a way, using Azure Blob Storage, to use IE to download any file directly from Azure Blob Storage?
    Thanks in advance.

    Hi,
    Actually, we can't set Content-Disposition for blobs, and I can't think of any other workarounds. From my experience, in most case IE's behavior is fine. I would like to know why you have to prompt a download? The user can see the text file, and
    if they wish to save it locally, they have more than one way to do that (copy paste, save file, etc.). If they simply want to read the text and then forget it, that's also fine. They don't even have to download it and then double click a local file to read
    the content.
    If you have to modify the behavior, the only workaround I can think of is to use a web role as an intermediate bridge, and add the Content-Disposition from your web role.
    Best Regards,
    Ming Xu.
    Please mark the replies as answers if they help or unmark if not.
    If you have any feedback about my replies, please contact
    [email protected]
    Microsoft One Code Framework

  • How to Secure or mask the Video Streaming on Azure Blob Storage

    I have a problem on using my Azure Blob Storage, this is for streaming specifically MP4 files.
    I have a wordpress site and I use Azure Blob Storage for my Storage, I use CloudberryExplorer for my browsing and editting
    My problem is that, the wordpress site Video Option which is MediaElemtJs have an exposed Video Source (This mean they can use the video source for their own project)
    Is their a way to mask or hide the video source? or I can limit the video source that it should only be coming from my site

    hi Franz,
    Base on my experience, there has two approaches for this issue.
    1.Using the SAS in js request.
    Like this:
    <source src="https://MY-AZURE-STORAGE.blob.core.windows.net/asset-b1ebfc63-4e1f-4c76-b2dd-b040113aa493/MigratingWebSitesAndDatabasesToWindowsAzure_mid.mp4?sv=2012-02-12&st=2013-09-03T17%3A17%3A20Z&se=2015-09-03T17%3A17%3A20Z&sr=c&si=efc57c3c-a025-41cd-a8b1-908231c7aed2&sig=xyiIxmqNdF%2Bo1x7eq89xCUgXkenZLEEpq%2B%2Bo715dT0U%3D" type="video/mp4">
    Please see this blog:http://azure.microsoft.com/en-us/documentation/articles/storage-dotnet-shared-access-signature-part-1/
    You could set the start time,expire time  and permission into your request. And you could use javascript to create the request and set src value. please see this sample (http://gauravmantri.com/2013/02/16/uploading-large-files-in-windows-azure-blob-storage-using-shared-access-signature-html-and-javascript/
    2.Create a web handler to handle the request url
    You could create a .ashx page and handle the url in .ashx page. Like this:
    <video src="videos.ashx?source=file.mp4" width="640" height="480" controls></video>
    Please try it.
    Also, you could see this threads
    http://stackoverflow.com/questions/9756837/prevent-html5-video-from-being-downloaded-right-click-saved form some ideas.
    Regards,
    Will 
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Large Files from FTP to Azure Storage Account.

             We required to transfer files from FTP to Azure Storage Account.
    Our FTP contains Almost 4 TB data and in future it will be growing.
    We require some automate process which will transfer data from FTP to Windows Azure Storage.
    Currently Implemented Solution:  We have created
    a windows service which will download files from FTP and converting to file stream and with Blob Storage object it will be uploaded to
    Storage account.
    With the mentioned approach bandwidth is utilized at 2 places
    For downloading file from FTP
    For Uploading to Azure Storage 
    Can you please provide any solution/suggestion which will reduce the consumption of bandwidth and large data will transfer in less time

    Hi,
    Please have a look at below article, it is talking about Blob Transfer Utility tool. Blob Transfer Utility is a GUI tool to upload and download thousands of small/large files to/from Windows Azure Blob Storage.
    #https://blobtransferutility.codeplex.com/
    Disclaimer: This is not a Microsoft official tool and it is not supported by Microsoft. It's just a (very useful) sample.
    Best Regards,
    Jambor
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Blob storage as output

    Hi All,
    I just started playing with the stream analytics to build a proof of concept app for our analytics. We're developing a widget that has millions of "loads" each day, these are the events we want to have analytics for.
    Due to the amount of data I'm trying to store the data in blob instead of sql. I've managed to configure the output to use blob storage. Is there a way to configure output to create a new blob every hour/day or similar? 
    BR,
    Max.

    Hi,
    I guess I can have a scheduled job running in my worker role that uses the management rest api to execute the following steps at predefined intervals.
    1. Stop the job
    2. Update the name of the job output file
    3. Start the job again using a custom time equal to when the job was stopped.
    Do you see any issues with this solution until support for partitioning the output is available?
    BR, Max.

  • Getting issue while downloading 4 to 5 gb file from blob storage.

    Hi i have to download a file from blob storage upto 3 to 5 gb .
    Below is my code:
    Label lblfilename = (Label)row.FindControl("lblGrid_filename");
        Label lblfilesize = (Label)row.FindControl("lblGrid_hidfileSize");
        string downloadfile = lblfilename.Text.ToString();
       // DownloadFileFromBlob(downloadfile, CONTAINER, ACCOUNTKEY);
        AccountFileTransfer = CloudStorageAccount.Parse("DefaultEndpointsProtocol=https;AccountName=" + ACCOUNTNAME + ";AccountKey=" + ACCOUNTKEY);
        if (AccountFileTransfer != null)
            BlobClientFileTransfer = AccountFileTransfer.CreateCloudBlobClient();
            ContainerFileTransfer = BlobClientFileTransfer.GetContainerReference(CONTAINER);
            ContainerFileTransfer.CreateIfNotExist();
            BlobRequestOptions options = new BlobRequestOptions();
            options.Timeout = new TimeSpan(0, 180, 0);
        var data = Regex.Match(lblfilesize.Text.ToString(), @"\d+").Value;
        int value32;
        Int64 value64;
        var blobSize = 0L;
        var blockSize = 0L;
        var offset = 0L;
        if (int.TryParse(data, out value32))
            blobSize = Convert.ToInt32(lblfilesize.Text.ToString());
        else if (Int64.TryParse(data, out value64))
            blobSize = Convert.ToInt64(lblfilesize.Text.ToString());
        var blob = ContainerFileTransfer.GetBlockBlobReference(downloadfile);
        var sasUrl = blob.Uri.AbsoluteUri;
        CloudBlockBlob blockBlob = new CloudBlockBlob(sasUrl);
        //blobSize = Convert.ToInt32(lblfilesize.Text.ToString());
        blockSize = 5 * 1024 * 1024;
        Response.Clear();
        Response.ContentType = "APPLICATION/OCTET-STREAM";
        System.String disHeader = "Attachment; Filename=\"" + blockBlob.Name + "\"";
        //Response.AppendHeader("Content-Disposition", disHeader);
        for (offset = 0; offset < blobSize; offset += blockSize)
            using (var blobStream = blockBlob.OpenRead())
                if ((offset + blockSize) > blobSize)
                    blockSize = (blobSize - offset);
                byte[] buffer = new byte[blockSize];
                blobStream.Read(buffer, 0, buffer.Length);
                Response.BinaryWrite(buffer);
                Response.Flush();
        Response.End();
    SNAPSHOT of error:
    Pawan

    Hi Will thanks per the reply
    I have make some changes in the code and now i am using the v4.0.30319 azure storage for downloading and able to download a file upto 3.5 GB. Now the problem is that each file size is showing 0 KB.
    Here is my code and snapshot:
      Button btndownloadrow = (Button)sender;
        GridViewRow row = (GridViewRow)btndownloadrow.NamingContainer;
        Label lblfilename = (Label)row.FindControl("lblGrid_filename");
        Label lblfilesize = (Label)row.FindControl("lblGrid_hidfileSize");
        string downloadfile = lblfilename.Text.ToString();
        CloudStorageAccount AccountFileTransfer = CloudStorageAccount.Parse("DefaultEndpointsProtocol=http;AccountName=" + ACCOUNTNAME + ";AccountKey=" + ACCOUNTKEY);
        // Create the blob client.
        CloudBlobClient BlobClientFileTransfer = AccountFileTransfer.CreateCloudBlobClient();
        // Retrieve reference to a previously created container.
        CloudBlobContainer ContainerFileTransfer = BlobClientFileTransfer.GetContainerReference("filetransfer");
        if (AccountFileTransfer != null)
            BlobClientFileTransfer = AccountFileTransfer.CreateCloudBlobClient();
            CloudBlockBlob pageBlob = ContainerFileTransfer.GetBlockBlobReference(CONTAINER);
            ContainerFileTransfer.CreateIfNotExists();
            BlobRequestOptions options = new BlobRequestOptions();
            options.ServerTimeout= new TimeSpan(0, 180, 0);
        var data = Regex.Match(lblfilesize.Text.ToString(), @"\d+").Value;
        int value32;
        Int64 value64;
        var blobSize = 0L;
        var blockSize = 0L;
        var offset = 0L;
        if (int.TryParse(data, out value32))
            blobSize = Convert.ToInt32(lblfilesize.Text.ToString());
        else if (Int64.TryParse(data, out value64))
            blobSize = Convert.ToInt64(lblfilesize.Text.ToString());
        CloudBlockBlob blob = ContainerFileTransfer.GetBlockBlobReference(downloadfile);
        var sasUrl = blob.Uri;
        CloudBlockBlob blockBlob = new CloudBlockBlob(sasUrl);
        //blobSize = Convert.ToInt32(lblfilesize.Text.ToString());
        blockSize = 4 * 1024 * 1024;
        Response.Clear();
        Response.ContentType = "APPLICATION/OCTET-STREAM";
        System.String disHeader = "Attachment; Filename=\"" + blockBlob.Name + "\"";
        Response.AppendHeader("Content-Disposition", disHeader);
       // blob.DownloadToStream(Response.OutputStream);
       // byte[] bytes = new byte[blockSize];
        for (offset = 0; offset < blobSize; offset += blockSize)
            using (var blobStream = blockBlob.OpenRead())
                if ((offset + blockSize) > blobSize)
                    blockSize = (blobSize - offset);
            byte[] buffer = new byte[blockSize];
                blobStream.Read(buffer, 0, buffer.Length);
                Response.BinaryWrite(buffer);
                Response.Flush();
        Response.End();
    SNAPSOT:
    Please suggest me if i am missing any logic.
    Pawan

  • How the files are decreasing from RBS storage after some time i delete docs from document library, where shld I check the settings for this?

    HI
    I created a web application and configured  rbs storage ,
    I uploaded documents (for ex: 18 number)
     after uploaded  documents ,  and observed the RBS storage folder, number of files are increasing (39 number) , its working fine
    and also  I run the querys to check the rows increasing or not , (when I upload new doc rows are increased its fine also)
    select count(*) from alldocs
    Select count(*) from AllDocstreams
    but when I delete some docs from document library ,
    1)the doc itself deleted from document library
    2)when I check the rbs storage  there is no number of files are decreased(its still 39 files ), its same as  before doc deleted from doc library
    after some time I found the number of files are decreasing from RBS storage folder in sql server,
    here I want to know the how the files are decreasing from RBS storage  after some time, where  shld I check the settings for this? and how I control on it
    here how I know the settings for cleaning up orphan BLBOS how  these deleted BLOB
    adil

    1. WHen you delete the file from Sharepoint, it is still present in Recycle bin. This is a default setting in sharepoint for 30 days. Once files are deleted fro Recycle bin, it can also be delete from RBS
    2. There is a RBS cleanup job which deletes files from RBS. for more info check 
    http://mehuljamod.blogspot.in/2012/09/remote-blob-storage-maintainer-rbs.html

  • Memory Dump :FBL5N No roll storage space of length 1753104 available for internal storage.

    Dear Friends,
               I am getting below error during T-Code FBL3N Report. No roll storage space of length 1753104 available for internal storage.
    Our Server' RAM size is 24 GB,and existing parameters are as below
    ztta/roll_area: 6500000
    ztta/roll_extension:2000683008
    abap/heap_area_total:2000683008
    abap/heap_area_dia: 5000000000
    abap/heap_area_nondia: 2000683008
    em/initial_size_MB: 4092
    Please suggest Parameter to be change.
    Details Dump log ST22
    Runtime Errors         TSV_TNEW_BLOCKS_NO_ROLL_MEMORY
    Date and Time          04.03.2014 10:18:48
    |Short Text                                                                                        |
    |    No roll storage space of length 1753104 available for internal storage.                       |
    |What happened?                                                                                    |
    |    Each transaction requires some main memory space to process                                   |
    |    application data. If the operating system cannot provide any more                             |
    |    space, the transaction is terminated.                                                         |
    |What can you do?                                                                                  |
    |                                                                                                  |
    |    Try to find out (e.g. by targetted data selection) whether the                                |
    |    transaction will run with less main memory.                                                   |
    |                                                                                                  |
    |    If there is a temporary bottleneck, execute the transaction again.                            |
    |    -                                                                                             |
    |                                                                                                  |
    |    If the error persists, ask your system administrator to check the                             |
    |    following profile parameters:                                                                 |
    |                                                                                                  |
    |    o  ztta/roll_area            (1.000.000 - 15.000.000)                                         |
    |           Classic roll area per user and internal mode                                           |
    |           usual amount of roll area per user and internal mode                                   |
    |    o  ztta/roll_extension       (10.000.000 - 500.000.000)                                       |
    |           Amount of memory per user in extended memory (EM)                                      |
    |    o  abap/heap_area_total      (100.000.000 - 1.500.000.000)                                    |
    |           Amount of memory (malloc) for all users of an application                              |
    |           server. If several background processes are running on                                 |
    |           one server, temporary bottlenecks may occur.                                           |
    |           Of course, the amount of memory (in bytes) must also be                                |
    |           available on the machine (main memory or file system swap).                            |
    |           Caution:                                                                               |
    |           The operating system must be set up so that there is also                              |
    |           enough memory for each process. Usually, the maximum address                           |
    |           space is too small.                                                                    |
    |           Ask your hardware manufacturer or your competence center                               |
    |           about this.                                                                            |
    |           In this case, consult your hardware vendor                                             |
    |    abap/heap_area_dia:        (10.000.000 - 1.000.000.000)                                       |
    |           Restriction of memory allocated to the heap with malloc                                |
    |           for each dialog process.                                                               |
    |    Parameters for background processes:                                                          |
    |    abap/heap_area_nondia:        (10.000.000 - 1.000.000.000)                                    |
    |           Restriction of memory allocated to the heap with malloc                                |
    |           for each background process.                                                           |
    |    Other memory-relevant parameters are:                                                         |
    |    em/initial_size_MB:         (35-1200)                                                         |
    |           Extended memory area from which all users of an                                        |
    |           application server can satisfy their memory requirement.                               |
    |    Note which actions and input led to the error.                                                |
    |                                                                                                  |
    |    For further help in handling the problem, contact your SAP administrator                      |
    |    .                                                                                             |
    |                                                                                                  |
    |    You can use the ABAP dump analysis transaction ST22 to view and manage                        |
    |    termination messages, in particular for long term reference.                                  |
    |                                                                                                  |
    |Error analysis                                                                                    |
    |    The internal table "\CLASS=ZCL_IM__GTA_TAX_CODE\METHOD=IF_EX_FI_ITEMS_CH_DATA~C               |
    |    HANGE_ITEMS\DATA=GT_BSIK" could not be enlarged further.                                      |
    |                                                                                                  |
    |    Memory location: "Session memory"                                                             |
    |                                                                                                  |
    |    You attempted to extend the data structure for the management of the                          |
    |    memory blocks for table "\CLASS=ZCL_IM__GTA_TAX_CODE\METHOD=IF_EX_FI_ITEMS_CH_D               |
    |    ATA~CHANGE_ITEMS\DATA=GT_BSIK". However, the 1753104 bytes required for                       |
    |    this were no longer available in the specified memory area.                                   |
    |                                                                                                  |
    |    The amount of memory requested is no longer available.                                        |
    |How to correct the error                                                                          |
    |                                                                                                  |
    |    Try to decide by analysis whether this request is                                             |
    |    reasonable or whether there is a program error. You should pay                                |
    |    particular attention to the internal table entries listed below.                              |
    |                                                                                                  |
    |                                                                                                  |
    |                                                                                                  |
    |                                                                                                  |
    |                                                                                                  |
    |                                                                                                  |
    |    If the error occures in a non-modified SAP program, you may be able to                        |
    |    find an interim solution in an SAP Note.                                                      |
    |    If you have access to SAP Notes, carry out a search with the following                        |
    |    keywords:                                                                                     |
    |                                                                                                  |
    |    "TSV_TNEW_BLOCKS_NO_ROLL_MEMORY" " "                                                          |
    |    "ZCL_IM__GTA_TAX_CODE==========CP" or "ZCL_IM__GTA_TAX_CODE==========CM001"                   |
    |    "IF_EX_FI_ITEMS_CH_DATA~CHANGE_ITEMS"                                                         |
    |                                                                                                  |
    |    If you cannot solve the problem yourself and want to send an error                            |
    |    notification to SAP, include the following information:                                       |
    |                                                                                                  |
    |    1. The description of the current problem (short dump)                                        |
    |                                                                                                  |
    |       To save the description, choose "System->List->Save->Local File                            |
    |    (Unconverted)".                                                                               |
    |                                                                                                  |
    |    2. Corresponding system log                                                                   |
    |                                                                                                  |
    |       Display the system log by calling transaction SM21.                                        |
    |       Restrict the time interval to 10 minutes before and five minutes                           |
    |    after the short dump. Then choose "System->List->Save->Local File                             |
    |    (Unconverted)".                                                                               |
    |                                                                                                  |
    |    3. If the problem occurs in a problem of your own or a modified SAP                           |
    |    program: The source code of the program                                                       |
    |       In the editor, choose "Utilities->More                                                     |
    |    Utilities->Upload/Download->Download".                                                        |
    |                                                                                                  |
    |    4. Details about the conditions under which the error occurred or which                       |
    |    actions and input led to the error.                                                           |
    |                                                                                                  |
    |System environment                                                                                |
    |    SAP-Release 700                                                                               |
    |                                                                                                  |
    |    Application server... "NEWPRD1"                                                              |
    |    Network address...... "10.0.0.1"                                                              |
    |    Operating system..... "Windows NT"                                                            |
    |    Release.............. "6.0"                                                                   |
    |    Hardware type........ "16x AMD64 Level"                                                       |
    |    Character length.... 16 Bits                                                                  |
    |    Pointer length....... 64 Bits                                                                 |
    |    Work process number.. 19                                                                      |
    |    Shortdump setting.... "full"                                                                  |
    |                                                                                                  |
    |    Database server... "NEWPRD1"                                                                 |
    |    Database type..... "ORACLE"                                                                   |
    |    Database name..... "PRD"                                                                      |
    |    Database user ID.. "SAPSR3"                                                                   |
    |                                                                                                  |
    |    Terminal................. " "                                                                 |
    |                                                                                                  |
    |    Char.set.... "C"                                                                              |
    |                                                                                                  |
    |    SAP kernel....... 700                                                                         |
    |    created (date)... "Sep 17 2012 22:56:00"                                                      |
    |    create on........ "NT 5.2 3790 Service Pack 2 x86 MS VC++ 14.00"                              |
    |    Database version. "OCI_10201_SHARE (10.2.0.4.0) "                                             |
    |                                                                                                  |
    |    Patch level. 353                                                                              |
    |    Patch text.. " "                                                                              |
    |                                                                                                  |
    |    Database............. "ORACLE 10.1.0.*.*, ORACLE 10.2.0.*.*, ORACLE 11.2.*.*.*"               |
    |    SAP database version. 700                                                                     |
    |    Operating system..... "Windows NT 5.0, Windows NT 5.1, Windows NT 5.2, Windows                |
    |     NT 6.0, Windows NT 6.1, Windows NT 6.2"                                                      |
    |                                                                                                  |
    |    Memory consumption                                                                            |
    |    Roll.... 16192                                                                                |
    |    EM...... 1826770240                                                                           |
    |    Heap.... 0                                                                                    |
    |    Page.... 40960                                                                                |
    |    MM Used. 1815573536                                                                           |
    |    MM Free. 2717408                                                                              |
    |User and Transaction                                                                              |
    |                                                                                                  |
    |    Client.............. 500                                                                      |
    |    User................ 1651                                                                     |
    |    Language key........ "E"                                                                      |
    |    Transaction......... " "                                                                      |
    |    Transactions ID..... "ED4EA3E3AB0AF11DA318E61F131BC713"                                       |
    |                                                                                                  |
    |    Program............. "ZCL_IM__GTA_TAX_CODE==========CP"                                       |
    |    Screen.............. "SAPMSSY0 1000"                                                          |
    |    Screen line......... 6                                                                        |
    |Information on where terminated                                                                   |
    |    Termination occurred in the ABAP program "ZCL_IM__GTA_TAX_CODE==========CP" -                 |
    |     in "IF_EX_FI_ITEMS_CH_DATA~CHANGE_ITEMS".                                                    |
    |    The main program was "RFITEMGL ".                                                             |
    |                                                                                                  |
    |    In the source code you have the termination point in line 64                                  |
    |    of the (Include) program "ZCL_IM__GTA_TAX_CODE==========CM001".                               |
    |    The program "ZCL_IM__GTA_TAX_CODE==========CP" was started as a background job.               |
    |    Job Name....... "RFITEMGL"                                                                    |
    |    Job Initiator.. "BASIS"                                                                       |
    |    Job Number..... 09115400                                                                      |
    |Source Code Extract                                                                               |
    |Line |SourceCde                                                                                   |
    |   34|          MANDT TYPE BSEG-MANDT,                                                            |
    |   35|          BUKRS TYPE BSEG-BUKRS,                                                            |
    |   36|          BELNR TYPE BSEG-BELNR,                                                            |
    |   37|          GJAHR TYPE BSEG-GJAHR,                                                            |
    |   38|          BUZEI TYPE BSEG-BUZEI,                                                            |
    |   39|          QSSKZ TYPE BSEG-QSSKZ,                                                            |
    |   40|          KTOSL TYPE BSEG-KTOSL,                                                            |
    |   41|          HKONT TYPE BSEG-HKONT,                                                            |
    |   42|         END OF TY_BSEG,                                                                    |
    |   43|                                                                                            |
    |   44|         BEGIN OF TY_FINAL,                                                                 |
    |   45|           BUKRS TYPE BSEG-BUKRS,                                                           |
    |   46|           BELNR TYPE BSEG-BELNR,                                                           |
    |   47|           GJAHR TYPE BSEG-GJAHR,                                                           |
    |   48|         END OF TY_FINAL.                                                                   |
    |   49|                                                                                            |
    |   50|  DATA: CW_ITEMS TYPE RFPOSXEXT,                                                            |
    |   51|        GT_BSIK TYPE STANDARD TABLE OF TY_BSIK,                                             |
    |   52|        GW_BSIK TYPE TY_BSIK,                                                               |
    |   53|        GT_BSAK TYPE STANDARD TABLE OF TY_BSAK,                                             |
    |   54|        GW_BSAK TYPE TY_BSAK,                                                               |
    |   55|        GT_BSEG TYPE STANDARD TABLE OF TY_BSEG,                                             |
    |   56|        GW_BSEG TYPE TY_BSEG,                              &nb

    sap basis wrote:
    |Information on where terminated                                                                   |
    |    Termination occurred in the ABAP program "ZCL_IM__GTA_TAX_CODE==========CP" -                 |
    |     in "IF_EX_FI_ITEMS_CH_DATA~CHANGE_ITEMS".                                                    |
    |    The main program was "RFITEMGL ".                                                             |
    |                                                                                                  |
    |    In the source code you have the termination point in line 64                                  |
    |    of the (Include) program "ZCL_IM__GTA_TAX_CODE==========CM001".                               |
    |    The program "ZCL_IM__GTA_TAX_CODE==========CP" was started as a background job.               |
    |    Job Name....... "RFITEMGL"                                                                    |
    |    Job Initiator.. "BASIS"                                                                       |
    |    Job Number..... 09115400                                                                      |
    First of all try to optimize code.

  • Access blob storage files by specific domain. (Prevent hotlinking in Azure Blob Storage)

    Hi,
    My application deployed on azure, and I managed all my file to blob storage.
    When i created container with public permission then it accessible for all anonymous users. When i hit URL of file (blob) from different browser, then i will get that file.
    In Our application we have some important file and images that we don't want to expose. When we render HTML page then in <img> tag we define src="{blob file url}" when i mention this then public file are accessible, but same URL i copied
    and hit to anther browser then still it is visible. My requirement is my application domain only able to access that public file in blob storage.
    Amazon S3 which provide bucket policy where we define that for specific domain only file will accessible. see http://docs.aws.amazon.com/AmazonS3/latest/dev/example-bucket-policies.html
    Restricting Access to a Specific HTTP Referrer

    hi Prasad,
    Thanks for your post back.
    All of SAS and CORS could work, but not comprehensive.
    For your requirement, " My requirement is my application domain only able to access that public file in blob storage.", If you want to stop the other domain site access your blob, you may need set the CORS for your blob. When
    the origin domain of the request is checked against the domains listed for the
    AllowedOrigins element. If the origin domain is included in the list, or all domains are allowed with the wildcard character '*', then rules evaluation proceeds. If the origin domain is not included, then the request fails. So other domain didn't access
    your resource. You also try the Gaurav's blog:
    http://gauravmantri.com/2013/12/01/windows-azure-storage-and-cors-lets-have-some-fun/
    If you access CROS resource, you also need use SAS authenticated.
    However SAS means that you can grant a client limited permissions to your blobs, queues, or tables for a specified period of time and with a specified set of permissions, without having to share your account access keys. The SAS is a URI that encompasses
    in its query parameters all of the information necessary for authenticated access to a storage resource (http://azure.microsoft.com/en-us/documentation/articles/storage-dotnet-shared-access-signature-part-1/
    ).  So if your SAS URI is available and not expired ,this URI could be used to other domain site. I think you can try to test it.
    If I misunderstood, please let me know.
    Regards,
    Will
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Renaming files in blobs storage

    Hi,
    I want to rename files in blobs storage which are inside the container. I can't find any demo code. Please help
    Regards,
    Nitin

    Please note that rename operation is not natively supported in blob storage. So in order to rename blobs, you would need to perform 2 operations - Copy & Delete.
    So what you will do is copy the blob to a new name. Because copy operation is async, you must wait for the copy operation to finish. Once copy operation is completed successfully, you would then delete the blob.
    Hope this helps.

  • Azure blob storage can not load file

    I am trying to follow the MSDN tutorial on creating a container in blob storage and am getting the following error:
    Additional information: Could not load file or assembly 'Microsoft.Data.Services.Client, Version=5.6.0.0, Culture=neutral, PublicKeyToken=31bf3856ad364e35' or one of its dependencies. The located assembly's manifest definition does not match the assembly reference. (Exception from HRESULT: 0x80131040)
    I have tried numerous stack overflow suggestions such as this
    one , to no avail. I have also tried uninstalling Azure storage, and re-installing it as well. Just in case it is helpful here is my test case for Azure storage:
    public void TestCloudStorage()
    CloudStorageAccount storageAccount = CloudStorageAccount.Parse(ConfigurationManager.AppSettings["StorageConnectionString"]);
    CloudBlobClient blobClient = storageAccount.CreateCloudBlobClient();
    CloudBlobContainer container = blobClient.GetContainerReference("pictures");
    container.CreateIfNotExists();
    Thank you in advance for your help!

    HI
    I know what happend, that's the Nuget Package error.
    Azure Storage nuget package 3.0+ have this error.
    My solution is use Azure nuget package 2.1+
    1 delete all the dll reference for Azure storage (Or recreate a new project)
    2 input below command in package Manager console
    Install-Package
    WindowsAzure.Storage -Version 2.1.0.4
    That will resolve your problem absolutely .
    Aslo you can refer to other peoples solution.
    http://stackoverflow.com/questions/20457846/missing-microsoft-data-services-client-version-5-6-on-azure-websites
    My Blog
    Please use Make as Answer if my post solved your problem and use
    Vote As Helpful if a post was useful.

Maybe you are looking for