Authorizations BW (only load flat-file and update hierarchy)

Hello,
Could you please explain how can I create an authorization for the user to only load-file and update hierarchy at BW?
Thanks,
Adriana

Dear Adriana,
I am not really sure about the exact answer for your question. However you can check the following links. It may be helpful.
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/39f29890-0201-0010-1197-f0ed3a0d279f
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/659fa0a2-0a01-0010-b39c-8f92b19fbfea
Thanks,
Raj

Similar Messages

  • Loading Flat File and Temporary Table question

    I am new to ODI. Is there a method to load a flat file directly in to the target table and not have ODI create a temporary file? I am asking this in trying to find out how efficient ODI is in bulk loading a very large set of data.

    You can always modify the KM to skip the loading of the temporary table and load the target table directly. Another option would be to use the SUNOPSIS MEMORY ENGINE as your staging database. But, this option has some drawbacks, as it relies on available memory for the JVM and is not suitable for large datasets (which you mention you have).
    http://odiexperts.com/11g-oracle-data-integrator-%E2%80%93-part-711g-%E2%80%93-sunopsis-memory-engine/
    Regards,
    Michael Rainey
    Edited by: Michael R on Oct 15, 2012 2:44 PM

  • Loading different flat files and consolidate in one BEX

    I want to know in SAP BW how is the best way to load different flat files and loading them into a Consolitaded Infocubes o DSO to show in a Bex report. This files have the same estructure.

    you need to create a datasource with the exact structure of your files... then this datasource will be linked with your cube via transformation.... you need to define the characters, key figures and time character ..
    load file to datasource.. then to cube....then it will be available for reporting

  • Part 2: Flat files and Business Contents: Any issues with this scenario?

    I will appreciate some clarification on the some points made in response to my previous post "Flat files and Business Contents: Any issues with this scenario?"
    1.
    " ...you’d better analyze those cubes for data redundancy and presence of data you’ll never use. " I will appreciate some clarification on the type of analysis you are referring to. Examples will help.
    2.
    "If you want to combine several found IOs in your custom dataprovider, then again you must know (or figure out) relationships between these IOs." I will appreciate some clarification on the type of relationship you are referring to. Examples will help.
    3.
    I am a bit confused with "..include into ODS structure ALL fields required for the cube" but you also noted noted that "...except navigational attributes and chars and KFs that are going to be determined in TRs or URs."
    If you exclude ALL these, haven't you excluded all the fields you included in the ODS structure?
    4.
    "Consider carefully the ODS’ key fields selection. Their combination should not allow data aggregation that you don’t need."
    I may be missing the point here, I understand that you need to select the fields which will form the unique ID for the records in the ODS under the Key Field (please correct me if I am wrong with the purpose of the Key Field), but I don't understand the discussion of "aggregation" in the context.
    Thanks in advance

    Hallo
    I try to give some exaplanation based on the previous answer.
    1. Data redundancy - make sure you do not store the same information. does not make sense to have data redundanty across you Data Warehouse. this is also a cost. just sotre the same information one time if you get all what you need.
    2. whatwhever you build you dp, which consist of IO, you need to know with kind of relation (1:1 or 1:n - n:n and so on) exist between them. that will help you when you model you infoprovider. For example I would never pit togheter IO (n:m) in the same dimension if you expect an high number of cardinality. Sometime an IO can be an attribute of another one (depend on relation. For example
    Business Partner and his Address. Usually you have a relation 1:1, in this case address is an attribute of BParten and store it in the Masterdata instead then DP
    3. Sometime when you load from ODS to CUBE, you can fill some IO (which are in the infocube and not in the ODS)through ABAP routine in TR-Start Routine of Update Rule. Does not make sense to include these IO in the ODS as they are NULL or Blank (the deault value). This can happen when for example, you first load in the ODS (Price and Quantity) and then you calculate Sell price later (Price * Quantity). of course it could be doen also in the Bex. Depends on other factors (Performance - Loaidng -Sizing)
    4. Make sure that the KEY definition of ODS is accordingly to the data otherwise you will aggregate the data and later maybe if you need the detail you miss it.
    for example: customer - product - Distr Chan - Sell Price
    if each Customer can buy each product for any Distrution Channel, then when you build your ODS(Customer - Product and Distribution must be KEY) otherwise (if you have only Customer - Product KEY for example) you will lose the details for Distribution Channel.
    I hope eveyrhting is clear
    Regards
    Mike

  • IDOC to flat file and vice versa

    Dear community,
    we should send IDOCs to flat files and read plain files and map them to IDOCs.
    Is there any comfortable way to do this without doing a huge content conversion for
    IDOCs? Exist a SAP standard for such process?
    We receive the IDOC as normal IDOC. Would it be more easy to save the IDOCs
    directly as flat file on ERP file system??
    System PI 7.11
    IDOCS: WHSCON, WHSORD and WMMBXY
    Regards
    Chris
    Edited by: Christian Riekenberg on Feb 1, 2012 2:41 PM

    Hey folks!
    We have a PI 7.31 SPS5 Java-only installation here and are using the module IDOCFlatToXmlConvertor. For single IDocs with one line for the EDC_DC40 header and the others for the IDoc payload the conversion works. But when we want to process multiple IDocs from the same type within one file, we see this error in the log:
    Module Exception 'com.sap.aii.af.idoc.exception.IDOCModuleException: No data segments for DOCNUM = ' found, cause: <null>
    Are there any requirements how the multiple IDocs have to be structured within the file?
    Best regards
    Heinrich
    Update:
    I found out that the module wants each record to contain the corresponding DOCNUM to find the matching records. The non-SAP sender system normally does not set this value. Is there a way to "convince" the module to convert the IDocs nevertheless?

  • Loading Flat File

    Hi Gurus, I have this scenario, where a ODS A (Auditing) has export datasource that feeds data to another ODS C (Auditing and Posting). ODS B (Posting) is independent and is not physically in the data flow.
    ODS C has update routines that does lookup for some some fields in ODS B. and eventually ODS C has update rules that feed data to CUBE A.
    So ODS C has fields from both ODS A and ODS B.
    I wanted to see if I can load a flat file into directly ODS C, it has a export datasource. I tried, I always get that famous 'InfoSource' error msg. Please guide.
    Please let me know if you need more explanations.
    Thanks.

    hi,
    it's possible to load flat file data directly to cube C. before loading the data to ods c. you have to decline the infoobject's in to ods. mention the flat file infoobjects in to ods. then do the loading. when you r creating the update rules from ods to infosource it's give some keyfigure has no update. ignore the error and activate the update rule.
    one thing i forget to say. when u r updating data from ods c to cube it will cause error, because of the addition of new object 's in to cube. it its show the error, delete the old export data source and recreate it newly.
    hope this help you
    regards
    harikrishna N

  • Risk factors while loading flat files

    HI,
    CAn anybody tell me what are risk factors while loading
    flat files....
    Thanks in advance,
    regards,
    Rags

    Risks:
    1) If flat-file is created manually
       - users may forget columns, specifically UNIT and
         CURRENCY.
       - Keyfigures my be presented in thousands one time
         and in single dollars the other time.
       - Flat-file may be delivered too late.
       - Automatic scheduling is difficult if flat-file
         comes from client station. You even can't do it
         with a process chain.
       - Naming conventions (capitals, leading zeroes,
         trailing spaces, special characters)
       - Master data conversion
    2) If flat-file is created automatically (from legacy
       system):
       - Data may arrive too late
       - Data file (if loaded from server) requires a
         connection to the external disk - security risk.
       - (Master) Data conversion
    Note: In the past I had trouble with uploading flat-files from a network-drive. It sometimes happened that the flatfile was only partly imported and that no errors where reported. When loading data from my local C-drive I never had this problem.
    My advise is to upload a flat-file into an ODS before using the data further in BW. In the ODS you can add a date to mark the loading data. If the loading to an ODS is ok then there is a much bigger change that you will not get problems further in the data-stream. Make sure that the key-user validates the data in the ODS before using it any further. It pays off to create one or two special queries on this ODS to enable testing of the data.

  • Loading Flat Files

    Hi,
    Iam trying to load flat files into BW. Does any one have a procedure for it. Iam also trying to figure out if we can automate the process if yes please give detail information about it. I would really appreciate if any one could answer to this.
    Thanks,
    Suneel

    Hi,
    If you have Unix authorizations on the app server you can do it yourself. You can create folders by a command called mkdir (not sure if this is the correct syntax). Its simple Unix code.Make sure that the id that is going to execute the job has read, write and execute authorization to these folders. You can take help of a Unix SA or Basis person for this.
    On the BW side, in teh flat file infopackage, you need to give this path for loading the file and then schedule the job as you do the other BW jobs.
    For extractions, you can put the destination path in whatever you are using for extraction, ex ABAP extracts, etc.
    For incoming and outgoing files, you can use SYNCRA or EFUSE or any other mechanism that is used to FTP files to get files into your app server or extract out files to downstream systems. You can again take the help of your Unix /BASIS people here.
    Cheers,
    Kedar

  • Loading flat files located on a server using the Control Center piece ofOWB

    Loading flat files from a server with OWB
    I am successful loading files with OWB, as long as they resided on my workstation in the C:\ drive. I copied the files to a Linux server into directories owned by Oracle/oinstall, permissions on the directories containing the files are 777. I ran the OWB client from the server with the following results:
    Error
    RPE-01013: SQL Loader reported error condition, number 1.
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 9 11:15:58 2007
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    SQL*Loader-500: Unable to open file (/u00/data/owb_repos1/state_codes.csv)
    SQL*Loader-553: file not found
    SQL*Loader-509: System error: The system cannot find the file specified.
    SQL*Loader: Release 10.2.0.1.0 - Production on Fri Mar 9 11:15:58 2007
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    SQL*Loader-500: Unable to open file (/u00/data/owb_repos1/state_codes.csv)
    SQL*Loader-553: file not found
    SQL*Loader-509: System error: The system cannot find the file specified.
    I tried to load the file using the control center from both the windows client and the server client. Can I do this from either one?
    I can't import the file in the Design Center from the /u00/data/owb_repos1 directory, it does not show up.
    If anyone has received similiar errors from OWB, please tell me how I can fix this. THANKS!!!
    [d97886@owbtest owb_repos1]$ pwd
    /u00/data/owb_repos1
    [d97886@owbtest owb_repos1]$ ls -la
    total 24
    drwxrwxrwx 2 oracle oinstall 4096 Mar 19 12:14 .
    drwxrwxrwx 18 oracle oinstall 4096 Mar 6 16:41 ..
    -rwxrwxrwx 1 oracle garrett 1481 Mar 8 07:32 state_codes.csv
    [d97886@owbtest owb_repos1]$

    Hi
    This is what i do for my flat files (i am loading my tables from the flat files):
    1) Create location for the files in OWB and it should look something like \\path_to_where_the_files_are. Here you should be able to browse to your Linux box and select the location of the files.
    2) Then check whatever you use (i use TOAD) and you should see in owb schema a directory created there which points to the location of the flat file.
    3) then just follow the guide to sample the files etc and you are good to go.
    Basically OWB has to be able to see the files that are on Linux machine.
    Alternatively you can map your Linux share to your windows box (use samba or something) and use it that way.
    Sorry if it is stating the obvious.
    Hope this helps
    Kind Regards
    Vix

  • Load Flat File from App Server

    HI, all.
    I want load flat file from appliation server. I created CSV file and loaded it on app. server with FM ARCHIVEFILE_CLIENT_TO_SERVER. After I created datasource and tried load data in the infocube. In this procedure I encountered with two problem
    1. When I look file in AL11, I doesn't see cyrillic symbols, instead this symbols I see #.
    2. When I try load data with data source, I get exception RS_EXCEPTION 000 "File don't open"
    Anybody can help me resolve this problem?
    wbr, Fanil.

    Hi, kodanda pani KV.
    2. File is was closed.
    1. Can me you clearly explain what you meen?
    wbr, Fanil

  • Getting error while loading Flat File

    Hello All,
    I am getting error while loading flat file. Flat file is a CSV file.
    Value ',,,,,,,,' of characteristic 0DATE is not a number with 000008 spaces
    Data seprator  |
    Escape sign    ;
    It has 23708 entries , it s loading successfully till 23 665 entries
    Besides when checked in PSA
    for record having entries >23667 has calender day as ,,,,,, where as rest entries are  having date
    Besides when i checked in Flat file ,the total number of rows is 23,667 is there but i wonder why it has got 23,708 in
    RSMO
    Could you please let me know how to correct.
    regards
    path

    Hi,
    For date column you should maintain YYYYMMDD formate Eg: 20090601, kepp cursor on date column and right click and Formate >Custome>make it 00000000 then save teh file as .CSV . First type values on column and do like this formate and save it and without opening it load it. Once you open it you losw 00000000 formate you need to give again the same formate.
    Settings in Infopackage:
    Data Format  = CSV
    Data Separator = ,
    Escape Sign = ;
    http://help.sap.com/saphelp_nw2004s/helpdata/en/43/01ed2fe3811a77e10000000a422035/content.htm
    http://help.sap.com/saphelp_nw2004s/helpdata/en/80/1a6567e07211d2acb80000e829fbfe/content.htm
    Thanks
    Reddy

  • Issue in calling webdynpro for loading flat file in IP

    Hello,
    We have implemented the How-To load flat files in IP paper and we are facing an issue in that. The generated webdynpro runs on ABAP Application server but we need to call this from a Web template which runs on portal server.
    These two servers are different and we don't want to hardcode the url in webtemplate as the server names will change in Q & P for webdynpro. Can anyone advise how we can read the ABAP Application server details in WAD may be using javascript and then generate the URL for webdynpro dynamically.
    Regards,
    Deepti

    Hi Deepti,
    We have also facing the same challenge. Could you please post the solution in which URL path change dynamically.
    I have created the ABAP function module based on the below link.
    http://wiki.sdn.sap.com/wiki/display/Snippets/BICustomWebItemtoLaunchanABAPWeb+Dynpro
    But not sure how to call the function module using custom web item & java script. Coould you please explain the steps on the same ?
    Thanks,
    Gopi R

  • Error handling when loading flate files

    Hi
    We have a Process Chain loading flate files from a server into a ODS. We want to create a routine to avoid errors in the ODS when no files is found.
    The error we want to handle is
    Error 1 when loading external data
    Message no. RSAR234
    We want to handle this error message as OK to avoid red request in the ODS
    Please help
    Kind regards
    Ola Einar Langen

    Hi, and thanks for your hints.
    We ended up with the following codes. It's will mark the prosess chain red (but this is as wanted) but we have implemented that the prosess chain will proceed even with error in the loading process
    The code for the routine implemented in the infopackage in "External data" tab.
    p_filename = 'filename.txt.
    OPEN DATASET p_filename FOR INPUT IN BINARY MODE.
    If file could not be open
    IF SY-SUBRC NE 0.
        p_subrc = SY-SUBRC.
      delete the current request
        CALL FUNCTION 'RSSM_DELETE_REQUEST'.
    ENDIF.
    close the file if OK
    CLOSE DATASET p_filename.
    p_subrc = 0.
    The solution is no OK for us
    Kind regards
    Ola Einar Langen

  • How to load flat file in BPS through Web-debugging

    Hi,
    We are working on flat file upload in BPS thru guide 'How to load flat file in BPS through Web'. Can some one guide on how to debug the function modules used while uploading the data.
    We have set up the break point in the function modules and also in the BSP page. But when trying to debug while uploading, it is not going to the break point which was set.
    Could you assist in setting up the break point.
    Regards,
    Sreenath
    Message was edited by:
            sreenath reddy

    Hi,
    I  have put an external break-point.
    I have hard coded it in the coding of the function module itself.
    The code is perfectly working. But, when I want to check for the values of the variables in the F.M during runtime, the break-point is not triggering. Any ideas??
    Regards,

  • I want rsync only download new files and keep old files

    I install Arch i686 in home and in home I do not have internet connection , so at work I download all packages I need with this command
    rsync -avrt --bwlimit=58 rsync://mirror.datacenter.by/arch//extra/os/any/ /home/ddd/i686/
    and move them to home and then copy them to
    /var/cache/pacman/arch
    and install packages with pacman
    after two days , I run this command
    rsync -avrt --bwlimit=58 rsync://mirror.datacenter.by/arch//extra/os/any/ /home/ddd/i686/
    and I see rsync get last packages again and use many Bandwidth
    how I tell rsync only download new files and keep old files

    rsync manpage says the following:
    -u, --update                skip files that are newer on the receiver
          --inplace               update destination files in-place
          --append                append data onto shorter files
          --append-verify         --append w/old data in file checksum
    I guess you need to add -u flag.
    Reading manpages is the highest virtue in GNU/Linux.

Maybe you are looking for