Syndicating Attributes in a flat file

Hello, I have a challenge.
I am want to syndicate a material classification (taxonomy table) and related attributes to a flat file. I want the result to look like this:
C 613316   VALVE_MANUAL_                 
D 613316   PRESSURE_RATING_CLASS_INLET_   Class 150                    
D 613316   MATERIAL_DESCRIPTION_BODY_     Duplex stainless steel       
D 613316   OPERATOR_TYPE_                 Manual                       
D 613316   CONNECTION_DESIGN_INLET_       RF                           
D 613316   VDS_REF_                       L-SP-200                       
D 613316   NOMINAL_SIZE_INLET_            1/2"                         
D 613316   VALVE_DATA_SHEET_              GT078               
C 613437   VALVE_MANUAL_                 
D 613437   MATERIAL_DESCRIPTION_BODY_     Duplex stainless steel       
D 613437   OPERATOR_TYPE_                 Manual                       
D 613437   CONNECTION_DESIGN_INLET_       RF                           
D 613437   PRESSURE_RATING_CLASS_INLET_   Class 150                    
D 613437   NOMINAL_SIZE_INLET_            3/4"                         
D 613437   VALVE_DATA_SHEET_              GT078               
D 613437   VDS_REF_                       L-SP-200          
1st column is C for class or D for attribute
2nd column  is material number (as you see one line per attribute).
3rd column is attribute name
4th column is attribute value.
Does anyone know a way to syndicate to a flat file like this?
Best regards
John-Kjell

Hi ,
Have not worked on taxonomy tables but to syndicate to a flat file , in the destination items tab - create the required fields in your case attr ,mat no ,attr name & attribute value ; map these to the source items and create a map of type flat and  maintain the item format as variable width , and the output will be as required.
Do let us know if your issue is resolved.
Regards,
Anita

Similar Messages

  • Error while uploading master data attributes thru flat file

    while uploading master data attributes thru flat file i am facing the below errow. can nyone let me know plz.
    Error 8 when compiling the upload program: row 227, message: Data type /BIC/CCDEIO_MAT5_ATTR was found in a new

    This just means that you changed the transfer rules/transfer structure and have not refreshed the workbench yet.  The infopackage data hasn't been updated ... this is similar to a datasource update, where you need to replicate to make sure BW knows the latest version. 
    The best way to correct this error is to reload RSA1.  Every time I make a change to the tr/ts, I always run /nrsa1 to refresh the workbench. 
    Brian

  • How to design Flat file for loading attribute dimension in a planning application

    Dear Gurus,
    I have a requirement to extract attribute dimensions from an essbase application and load it to another planning application. I have a dimension called Program and two attribute dimensions Sales Manager, Accounts manager associated with Program dimension in Essbase application. I will Extract these dimensions using Essbase outline extractor. After Extracting the attribute dimensions I have to load these dimensions to planning applications using outline load utility. Kindly guide me how to design the flat file for loading attribute dimensions in planning application.
    Thanks and Regards
    SC

    You could dig through the docs and try to figure out the file format manually, or you could do this the easy way.  Simply use the Outline Load Utility to export your attribute dimension from Planning.  The export file format is the same as the import file format.  You might have to manually add a couple of test members to your attribute dimension so that your export file has some content.  Then simply update the file you exported, and import it.
    (I am assuming you have already manually created the Attribute dimension in Planning, and that you simply need to add members to it.)
    Hope this helps,
    - Jake

  • APD - Query output on to a Flat file

    All
    We are on BI 7.0 with SP13 and NWS 12. I am trying to create a new APD with source as "Query" built on Info object and target as "Flat File". There is no requirement of transformation so I am trying to map directly from Source to Target with a direct link with out using any transformations in between,
    I could check and found no errors in APD definition and when I run APD, the file is getting created both on application server and local workstation but with 0 records.
    The volume of records are expected to be around 4 millions on this query and am just wondering why APD is not writing any data on CSV file.
    Pls let me know the following;
    - What is wrong in the above process and why it is not writing data on to flat file
    - Can APD handle volume of 4 million rows even though job runs in the background?
    Your inputs in this are highly helpful and thanks for your time to provide solution.
    Thanks - Sriram

    Hi
    Thanks for your reply and I could solve this issue and found the root cause.
    The very purpose of this APD is to provide a list of business partners to an upstream systerm with few attributes and in this process one of my consultant has deactivated the key figure in the query which is an input for APD process.
    When the key figure is inactivated ( this is the only key figure in the query ), APD does not recognize query as a valid one and sends out an error message.
    When we re activate the key figure ( I have hidden this later using APD transformations to hide the column and send output to a flat file on to unix server ) in the query, the APD runs fine and could deliver the output in flat file.
    Thanks

  • Extracting XML Data to a flat file

    I have an application which is puting an XML string into a field in the database... now my problem is that I have to extract that information and put it in either 1. a flat file, or 2. a seperate table containing a column for each XML attribute... Does anyone know of the best way to approach either solution?
    cheers.

    And your Oracle version is?
    There are many different ways to put something into a file. How about UTL_FILE?http://www.psoug.org/reference/utl_file.html
    How to get it out of the table in the first place?
    I usually try a SELECT statement. They seem to work.
    PS: Databases don't have fields ... table's don't either. Check your verbiage.
    Try looking of the word "FIELD" in the Oracle docs. Then try "COLUMN."

  • How to Extract data to a Flat File?

    Hello,
    Im using OWB 10g and Oracle 9.2 version of database. I want to know the complete steps of creating the flat file as my target module. I need to extract the data (source) from the oracle tables and load it into the flat files. Pls let me know your suggestions. Pls note except the flat files, I do not have any oracle tables as my targets.
    Thanks,
    Kishan

    Hi, here's an excerpt from some notes I wrote a few months ago, they shpoud be clear enough.
    0) the file will be created on the server using the utl_file package. So the path where the file will be created must be added to the utl_file_dir Oracle Server parameter. Request your DBA to complete this step.
    -In OWB client:
    1) create a file location (or use existing one) under "Files/Locations" in project tree
    2) create a new Flat File module (or use existing one) under "Files" ; associate it with the desired location (not really necessary)
    3) create a new mapping in an Oracle module
    4) add to mapping a table operator bound to the table to dump
    5) add to mapping un unbound Flat File (FF) Operator, give it a meaningful name
    6) drag the desired attributes from table to FF Operator. This will instruct OWB to feed the file with comma-separated values from the source table using the utl_file package. Be careful to format attributes accordingly to the format you want.
    7) reconcile outbound the FF Operator, creating a new flat file in the module created during step 2
    8) close mapping, right-click on it an from menu "Configure..."->"Sources&Targets"-><YOURTARGETFILE>, specify the desired output location and the output file name
    9) Open deployment manager and connect to runtime rep. If the FF location is new, register it by right-clicking+menu "Register" and specify parameters: operating system user/passawrd of the DB server, host and path where the file will be stored (must be the same added to utl_file_dir). ATTENTION: path info is stored in the generated mapping so be sure to redeploy mappings if changes occur
    10) deploy mapping, test execution
    11) OPTIONAL you can specify a different delimiter (TAB,etc) for the target file directly in its editor. But be careful then to reconcile inbound the FF op in the mapping before deploying.
    Cheers, Antonio

  • Does any one know how to resolve the error while flat file upload

    Hey guys, I am so struck with this problem, please do let me know if anyone has any suggestions and ideas.
    1. I created three InfoObjects, Material Number, Name and Price,which is  a keyfigure and has data type DEC and fixed currency USD.
    Material Number is "With Master Data ", "With Text" and i added Name and Price as attributes to that.
    2.
    After that i created an Application component, and therafter a data source for Master data Attibutes.
    Now i select my flat file which is .csv file, in this data source. When I click "Load Example Data", i get all the fields from flat file, but the data type for PRICE changes to FLTP with length 16 instead of remaining DEC.
    Does anyone have any idea why on earth is this change of Data type taking place ?
    Please do me a favour by answering if you know about this problem. Thanks, appreciate it . Byee. And yeah, m just starting off in SAP so I am a total amateur unlike most of the people here.

    Am workin on BI 7.0.
    Let me put it in another way :
    Created Infoobjects:      1) Material No , 2) Material Name  3 ) Price which is the Key Figure
    Material No -
    has attributes Material Name and Price and " WIth Master Data", With ' Text"
    Price: -
    data type DEC, fixed currency - USD
    After doing things.
    Created Application Component:
    then, Right Click - Create Data Source,
    Selected -- Master Data Attributes
    After that from "EXTRACTIO' tab , selected the flat file, which is *.csv
    and then in DATA FORMAT - " SEPERATED WITH SEPERATOR"
    Then from "PROPOSAL" tab, when i click Load example data, the data from flat file shows up here, but the field PRICE now has Data Type -- " FLTP" and lenght "16"
    I hope i was clear enough. thanks

  • Loading Master Data from a Flat File

    Hi Guru, I have flat files of different SAP tables (Transactional data tables and Master Data ), I want to upload this flat files to a DSO, my question is in the case of Master Data Files example SAP table USR03 ( User Address Data) in a Flat file format which is better to create a Master data text and attributes Datasource or create it as a transactional data Datasource to upload it to a DSO.

    Hi
    Master data datasource should be used to load master data like customer details. customer data is the master data which does not change frequently. customer infoobject may have some attributes like customer name, customer contact number, customer address, etc. we have to maintain these attributes and text data.
    Transaction data is something which keeps changing very frequently like price of the material, sales, etc. for loading such data we go for transaction datasources.
    As mentioned above too,the changes in the address are very rare, therefore youshould not use transactional datasource.
    Regards, Rahul
    Edited by: Rahul Pant on Feb 7, 2011 12:24 PM

  • OIM 11.1.1.5 Flat File Reconciliation :

    Hi,
    I am using a csv file as a Target Source and trying to reconcile in OIM. I can successfully do so unless I try to populate the manager field in OIM. As soon as I try to populate the Manager field, I get an error.
    I read that the Manager attribute in OIM User Form is mapped to the usr_mgr_key in the backend. I believe it should automatically get the usr_key and map that to the usr_mgr_key upon reconciliation.
    In my example : Supervisor_ID = RANDY_MARSH (with user key is 61 already) exists in my OIM.
    The error I am getting is as follows : Any input is highly appreciated.
    ++usrColumnsList = usr_manager_key,usr_type,usr_title,usr_emp_no,usr_email,usr_dept_no,usr_middle_name,usr_login,act_key,usr_last_name,usr_first_name++
    ++usrDataArray = [61, End-User, Non Billable, 100002, [email protected], 4168500121, Theodore, 100002, 24, Cartman, Eric]++
    ++matchingRulesWhereClause = (((USR.USR_EMP_NO='100002')AND(USR.USR_TYPE='End-User')AND(USR.USR_DEPT_NO='4168500121')AND(USR.USR_LAST_NAME='Cartman')AND(USR.USR_TITLE='Non Billable')AND(UPPER(USRMGR.USR_LOGIN)=UPPER('61'))AND(USR.ACT_KEY='24')AND(USR.USR_EMAIL='[email protected]')AND(UPPER(USR.USR_LOGIN)=UPPER('100002'))AND(USR.USR_FIRST_NAME='Eric')AND(USR.USR_MIDDLE_NAME='Theodore')))++
    ++dateFormat = MM-DD-RRRR HH24:MI:SS++
    ++]]++
    ++[2011-12-28T16:33:30.203+05:30] [oim_server1] [ERROR] [IAM-5010000] [oracle.iam.reconciliation.impl] [tid: OIMQuartzScheduler_Worker-2] [userId: oiminternal] [ecid: d0b6053d8947039f:239a101:1345ff4d6f6:-8000-0000000000000002,0] [APP: oim#11.1.1.3.0] Generic Information: {0}[[++
    ++oracle.iam.platform.utils.SuperRuntimeException: java.sql.SQLSyntaxErrorException: ORA-00904: "USRMGR"."USR_LOGIN": invalid identifier++
    ++ORA-06512: at "DEV_OIM.XL_SP_RECON_USR_MATCH", line 14++
    ++ORA-06512: at "DEV_OIM.XL_SP_RECON_USR_DATA_MATCH", line 31++
    ++ORA-06512: at line 1++
    ++     at oracle.iam.reconciliation.dao.event.EventMgmtDao.ignoreEventUsrDataMatch(EventMgmtDao.java:450)++
    ++     at oracle.iam.reconciliation.impl.ReconOperationsServiceImpl.ignoreEvent(ReconOperationsServiceImpl.java:385)++
    ++     at oracle.iam.reconciliation.impl.ReconOperationsServiceImpl.ignoreEvent(ReconOperationsServiceImpl.java:346)++
    ++     at Thor.API.Operations.tcReconciliationOperationsIntfEJB.ignoreEventx(Unknown Source)++
    ++     at sun.reflect.GeneratedMethodAccessor9086.invoke(Unknown Source)++
    ++     at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)++
    ++     at java.lang.reflect.Method.invoke(Method.java:597)++

    I tried passing 61 but it didnt work.
    Anyways I was able to troubleshoot the error. It was rather a mistake on my side. There was a required field that I was missing on this flat file, EmployeeType that maps to Role on the OIM side. However the error shown in the logs was a bit misleading.
    Thanks for you help.

  • Flat file to Oracle Database - Invalid Month

    Hi All
    I am New to XI and trying to create a simple interface between Legacy system (flat file) and Oracle database.
    I have some date fields in the Oracle table.
    When i run the interface it says "invalid month" with some SQL exception number.
    I am using (mm/dd/yyyy) format for the date.
    Is there any conversion i need to do. I am using XML SQL format for Oracle database.
    Any help is appreciated.
    Thanks
    Nagakishore

    Hello,
    Oracle is very picky with the way date is stored. Assume that you have a field datefield that you need to save. In mapping add an attribute hasqQuote and set it to "NO". When mapping the value to the node use to_date as shown below.
    <DateField hasQuot="No">
       TO_DATE("2004-07-20", "yyyy-mm-dd hh:mi:ss")</DateField>
    cheers,
    Naveen

  • Error in Performing Flat-File Reconciliation

    Hi All,
    I was trying perform flat file reconcilation using the GTC Connector. The format of my flat file (PPSLocal.txt) is like this:
    Account Name|Full Name|Domain|Last Logon Timestamp|Description|GUID|Mail|Employee ID|First Name|Last Name
    PPS\SophosSAUPPS010|SophosSAUPPS010|PPS.LOCAL||Used for download of Sophos updates|EED86D86-750C-404A-9326-044A4DB07477||||
    PPS\GBPPL-SI08$|GBPPL-SI08$|PPS.LOCAL|||79677F4D-8959-493E-9CF9-CDDDB175E40B||||
    PPS\S6Services|Series 6 Services Account|PPS.LOCAL||Series 6 Services Account|B4F41EE2-6744-4064-95F6-74E017D0B9AF||||
    I created a GTC Connector "OtherDomain" specified all the configuration;
    Staging Directory (Parent identity data) /home/GTC
    Archiving Directory /home/GTC/archive
    File Prefix PPS
    Specified Delimiter |
    File Encoding UTF8
    Source Date Format yyyy/MM/dd hh:mm:ss z
    Reconcile Deletion of Multivalued Attribute Data check box unchecked
    Reconciliation Type Full
    After that i performed the connector configuration mapping and ran the respective schedule task which displayed the following error message on console:
    Class/Medthod: SharedDriveReconTransportProvider/getFirstPage - Before calling: getAllData
    *DEBUG,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider/getAllData entered.*
    *DEBUG,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider/getReconFileList entered.*
    *DEBUG,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider.EndsWithFilter/accept entered.*
    *INFO,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Staging dir -->/home/GTC*
    *INFO,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Number of files available -->0*
    *DEBUG,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider/getReconFileList left.*
    *ERROR,20 Aug 2009 09:08:49,202,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Problem encountered in reconciling the first page of data*
    com.thortech.xl.gc.exception.ReconciliationTransportException: No Parent files in staging directory or necassary access(READ) permissions are missing
    *     at com.thortech.xl.gc.impl.recon.SharedDriveReconTransportProvider.getAllData(Unknown Source)*
    *     at com.thortech.xl.gc.impl.recon.SharedDriveReconTransportProvider.getFirstPage(Unknown Source)*
    *     at com.thortech.xl.gc.runtime.GCScheduleTask.execute(Unknown Source)*
    *     at com.thortech.xl.scheduler.tasks.SchedulerBaseTask.run(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper$TaskExecutionAction.run(Unknown Source)*
    *     at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)*
    *     at weblogic.security.service.SecurityManager.runAs(Unknown Source)*
    *     at weblogic.security.Security.runAs(Security.java:41)*
    *     at Thor.API.Security.LoginHandler.weblogicLoginSession.runAs(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper.execute(Unknown Source)*
    *     at org.quartz.core.JobRunShell.run(JobRunShell.java:178)*
    *     at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:477)*
    *ERROR,20 Aug 2009 09:08:49,202,[XELLERATE.GC.FRAMEWORKRECONCILIATION],Reconciliation Encountered error:*
    com.thortech.xl.gc.exception.ProviderException: No Parent files in staging directory or necassary access(READ) permissions are missing
    *     at com.thortech.xl.gc.impl.recon.SharedDriveReconTransportProvider.getFirstPage(Unknown Source)*
    *     at com.thortech.xl.gc.runtime.GCScheduleTask.execute(Unknown Source)*
    *     at com.thortech.xl.scheduler.tasks.SchedulerBaseTask.run(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper$TaskExecutionAction.run(Unknown Source)*
    *     at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)*
    *     at weblogic.security.service.SecurityManager.runAs(Unknown Source)*
    *     at weblogic.security.Security.runAs(Security.java:41)*
    *     at Thor.API.Security.LoginHandler.weblogicLoginSession.runAs(Unknown Source)*
    *     at com.thortech.xl.scheduler.core.quartz.QuartzWrapper.execute(Unknown Source)*
    *     at org.quartz.core.JobRunShell.run(JobRunShell.java:178)*
    *     at org.quartz.simpl.SimpleThreadPool$WorkerThread.run(SimpleThreadPool.java:477)*
    Caused by: com.thortech.xl.gc.exception.ReconciliationTransportException: No Parent files in staging directory or necassary access(READ) permissions are missing
    *     at com.thortech.xl.gc.impl.recon.SharedDriveReconTransportProvider.getAllData(Unknown Source)*
    *     ... 11 more*
    *WARN,20 Aug 2009 09:08:49,203,[XELLERATE.GC.FRAMEWORKRECONCILIATION],Though Reconciliation Scheduled task has encountered an error, Reconciliation Transport providers have been "ended" smoothly. Any provider operation that occurs during that "end" or "clean-up" phase would have been executed e.g. Data archival. In case you want that data to be a part of next Reconciliation execution, restore it from Staging. Provider logs must be containing details about storage entities that would have been archived*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider/end entered.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Medthod: SharedDriveReconTransportProvider/end - After calling: Re-setting instance variables*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Medthod: SharedDriveReconTransportProvider/end - After calling: Re-set over for instance variables*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.GC.PROVIDER.RECONCILIATIONTRANSPORT],Class/Method: SharedDriveReconTransportProvider/end left.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SCHEDULER.TASK],Class/Method: SchedulerBaseTask/run left.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SCHEDULER.TASK],Class/Method: SchedulerBaseTask/isSuccess entered.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SCHEDULER.TASK],Class/Method: SchedulerBaseTask/isSuccess left.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SERVER],Class/Method: SchedulerTaskLocater /removeLocalTask entered.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SERVER],Class/Method: SchedulerTaskLocater /removeLocalTask left.*
    *DEBUG,20 Aug 2009 09:08:49,203,[XELLERATE.SERVER],Class/Method: QuartzWrapper/updateStatusToInactive entered.*
    *DEBUG,20 Aug 2009 09:08:49,207,[XELLERATE.SERVER],Class/Method: QuartzWrapper/updateStatusToInactive left.*
    *DEBUG,20 Aug 2009 09:08:49,207,[XELLERATE.SERVER],Class/Method: QuartzWrapper/updateTaskHistory entered.*
    *DEBUG,20 Aug 2009 09:08:49,208,[XELLERATE.SERVER],Class/Method: QuartzWrapper/updateTaskHistory left.*
    *DEBUG,20 Aug 2009 09:08:49,209,[XELLERATE.SERVER],Clearing Security Associations with thread executing Scheduled task*
    *DEBUG,20 Aug 2009 09:08:49,210,[XELLERATE.SERVER],Class/Method: QuartzWrapper/run left.*
    *DEBUG,20 Aug 2009 09:08:49,210,[XELLERATE.SERVER],Class/Method: QuartzWrapper/execute left.*
    However, to my amazement, the reconciliation event for all the user are getting created in OIM Design console;
    I have tried changing the directory /home/GTC permission to 777 also but it didnt help.
    If anyone has any clue on this, kindly help.
    Cheers,
    Sunny

    AS you are saying that fiel is there.
    Can you please tell me how your recon sch task is running, I mean to say at what interval or once or daily ?
    If you are putting your file in parent form it is going to archive after recon. that';s correct and your recon manager is getting users that is also correct it means your recon si working fine.
    Now to run this recon second time you have to put one more fiel with new records or updated records in the parent dir otherwise it shows error.
    For me it is not showing any error.
    As user774847 is saying that it is known issue, i can't comment on this as I am not getting this error. I get this error only when i don't put any flat file in parent dir and i run the sch task.
    Just put new file or updated one for each time you run the SCh Task and make the SCh Task ONCE or Daily.
    Let me know if errors are still coming

  • How to determine the size of a flat file attachment in a mail sender cc?

    Hi
    I use PI 7.11
    I have a scenario where a flat file attachment is being picked up by a MailSender adapter and if the size of the attached flat file is larger that 500 bytes the receiver is A and if the attachment is less that 500 bytes the receiver is B.
    I have checked the Context Object list in the Conditions section of Receiver Determination, and it seems, that only the file adapter allows for validation on the file size.
    I contemplated an extended receiver determination but the attachment is a flat file which just needs to be passed thru without being converted to an XML document, so my source document is not of XML format.
    An other but not very nice way would be to use an intermediate folder to drop the attachment in and from there use a file sender adapter and thus get access to the filesize attribute in my receiver determination, but I would like to avoid this.
    Any suggestions?
    BR MIkael

    Hi
    I have decided to make a module, where I plan to place the size of the attachment in the Dynamic Configuration variable "ProcessStep", but to my surprise, it seems that all the context objects in the receiver determination condition section are of type String and hence it is not possible to make a condition where I test for whether the value is smaller than say 500!? Only equal, NOT equal and LIKE are available (EX seems to be disabled as this probably only works in an xpath.
    Would one way be to use the xpath expression like this ns1:Main\ProcessStep < 500 where ns1: http://sap.com/xi/XI/System?
    MIkael

  • What are the settings for datasource and infopackage for flat file loading

    hI
    Im trying to load the data from flat file to DSO . can anyone tel me what are the settings for datasource and infopackage for flat file loading .
    pls let me know
    regards
    kumar

    Loading of transaction data in BI 7.0:step by step guide on how to load data from a flatfile into the BI 7 system
    Uploading of Transaction data
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( Transaction data )
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to create ODS( Data store object ) or Cube.
    • Specify name fro the ODS or cube and click create
    • From the template window select the required characteristics and key figures and drag and drop it into the DATA FIELD and KEY FIELDS
    • Click Activate.
    • Right click on ODS or Cube and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.
    4. Monitor
    Right Click data targets and select manage and in contents tab select contents to view the loaded data. There are two tables in ODS new table and active table to load data from new table to active table you have to activate after selecting the loaded data . Alternatively monitor icon can be used.
    Loading of master data in BI 7.0:
    For Uploading of master data in BI 7.0
    Log on to your SAP
    Transaction code RSA1—LEAD YOU TO MODELLING
    1. Creation of Info Objects
    • In left panel select info object
    • Create info area
    • Create info object catalog ( characteristics & Key figures ) by right clicking the created info area
    • Create new characteristics and key figures under respective catalogs according to the project requirement
    • Create required info objects and Activate.
    2. Creation of Data Source
    • In the left panel select data sources
    • Create application component(AC)
    • Right click AC and create datasource
    • Specify data source name, source system, and data type ( master data attributes, text, hierarchies)
    • In general tab give short, medium, and long description.
    • In extraction tab specify file path, header rows to be ignored, data format(csv) and data separator( , )
    • In proposal tab load example data and verify it.
    • In field tab you can you can give the technical name of info objects in the template and you not have to map during the transformation the server will automatically map accordingly. If you are not mapping in this field tab you have to manually map during the transformation in Info providers.
    • Activate data source and read preview data under preview tab.
    • Create info package by right clicking data source and in schedule tab click star to load data to PSA.( make sure to close the flat file during loading )
    3. Creation of data targets
    • In left panel select info provider
    • Select created info area and right click to select Insert Characteristics as info provider
    • Select required info object ( Ex : Employee ID)
    • Under that info object select attributes
    • Right click on attributes and select create transformation.
    • In source of transformation , select object type( data source) and specify its name and source system Note: Source system will be a temporary folder or package into which data is getting stored
    • Activate created transformation
    • Create Data transfer process (DTP) by right clicking the master data attributes
    • In extraction tab specify extraction mode ( full)
    • In update tab specify error handling ( request green)
    • Activate DTP and in execute tab click execute button to load data in data targets.

  • Data Source creation for Master Data from Flat File to BW

    Hi,
    I need to upload Master Data from Flat File. Can anybody tell step by step right from begining of creation of DataSource up to Loading into Master Data InfoObject.
    can any body have Document.
    Regards,
    Chakri.

    Hi,
    This is the procedure.
    1. Create m-data with or without attributes.
    2. Create infosource .
        a) with flexible update
             or
        b) with direct update
    3. Create transfer rules and assign tyhe names of m-data and attribute in "Transfer rules" tab and transfer them to communication structure.
    4. Create the flat-file with same stucture as communication structure.
    5. if chosen direct update then create infopackage and assign the name of flat-file and schedule it.
    6. if chosen flexible update the create update rule with assigning name of the infosource and the schedule it by creating infopackage.
    Hope this helps. If still have prob then let me know.
    Follow this link also.
    http://help.sap.com/saphelp_nw2004s/helpdata/en/b2/e50138fede083de10000009b38f8cf/frameset.htm
    Assign points if helpful.
    Vinod.

  • Start Routine to load flat file - how to track filename in transformation

    Hi,
    Could you please help me to write a start routine to bring data from a flat file with a file name week45_c1_pri_2006, where the wekk name week-- changes per week. how to track the filename in the routine and how in data transfer process knows to load data from each flat file.
    Message was edited by:
            Anima M

    PROGRAM trans_routine.
          CLASS routine DEFINITION
    CLASS lcl_transform DEFINITION.
      PUBLIC SECTION.
    Attributs
        DATA:
          p_check_master_data_exist
                TYPE RSODSOCHECKONLY READ-ONLY,
    *-    Instance for getting request runtime attributs;
        Available information: Refer to methods of
        interface 'if_rsbk_request_admintab_view'
          <b>p_r_request
                TYPE REF TO if_rsbk_request_admintab_view READ-ONLY.</b>
    <i>The bold area above where you get the requestid, here I am stuck.
    The below code does not bring the correct requestid of PSA</i>
    CALL METHOD P_R_REQUEST->GET_REQUID30
          RECEIVING
            R_REQUID30 = IREQUID.
        CALL METHOD P_R_REQUEST->GET_REQUID
          RECEIVING
            R_REQUID = IREQUID.
        SELECT FILENAME INTO IFILENAME FROM rsseldone WHERE rnr =  IREQUID  AND lnr = '        1'.
    Hope this can give a better picture where I am currently stuck not able to get the PSA request id in the <b>start routine of transformation</b> to read the data and filename from there.
    Thanks alot
    Anima

Maybe you are looking for

  • Where can I get an HP Laser Jet 5 Basic Printer driver

    Where can I get an HP Laser Jet 5 Basic Printer driver, besides using the windows Update button. 

  • Internet Radio crashes every 10 min. after update

    There are a few questions in the Support database that match this behaviour, but they are from 2012, so here it is again ... Just updated AppleTV yesterday (Sept 20, 2014) to iOS 8 update, and now Internet Radio crashes (any and all streams I pick),

  • Lumia 710 video recording

    Just got my new Lumia 710 today. My last phone was a WP 6.5. The interface is very different. I tried video recording on the Lumia. I do not get any(or faint) sound recording during video recording. Can I confirm there is a mic at the rear of the pho

  • SAP BW 3.52 and POS DM: Architecture Best Practise

    We are retail/wholesale company that are two different business units. Currently we are running SAP BW 3.0 that is integrated with SAP R/3 and APO for reporting purposes. We are now in the early stages of implementing SAP Retail along with SAP POS DM

  • Stop Underlining in mac mail

    How do you stop underlining in MacMail. No matter how many times I press the button it just keeps underlining