AET and extraction into BI

Hi guys,  BI consultant here so apologies I don't speak the language but...
the CRM developer has added 3 fields to crmd_order_h via AET.  He has marked all the 3 fields as 'BW reporting relevant'.
I have now checked the BW extractors in the hope that these 3 fields will now be appended to the extract structure for the BW datasources.
The Service order header extractor (0CRM_SRV_PROCESS_H) has been appended correctly with the 3 fields.  But the extractor which I will be using (0CRM_SRV_REQ_INCI_H) for service requests has not been enhanced.  Both extractors are based on crmd_order_h but have different business transaction types.  The CRM developer did not specify which bus type to apply the 3 fields to so I am not sure why the extractor for one bus type has been enhanced but not another.
Any ideas?

Hi Guy,
it looks like the datasource 0CRM_SRV_REQ_INCI_H was either not intended to be extended by the responsible developers or just forgotten.
For the AET there is a registry, which tells what objects have to be generated for which flag (TA AXTREG). I have not found any entry for your datasource there.
If it is really required to extend it, you will have to raise a customer message and to report this issue.
Best regards
Matthias

Similar Messages

  • Custom infotype in HCM and Extraction into BI

    Hi
    I am unable to build a custom extractor from view in hcm to extract the infotype 0102.
    could you please help me out in understanding hcm custom extractions?
    The error message was: Invalid extract structure template PA0102.
    Can I not create a custom extractor over any infotype? what about the custom infotypes?
    Please help me in understanding this: If we create a custom infotype in R3, will all the data related to that reside in one underlying table or not?
    Can I create a custom extractor to extract from the custom infotype without ABAP?
    Thanks
    Sarah

    Hi Sarah,
    I guess your is related to some of the amount or quantity fields included in pa0102. Check it out. They must point to unit fields also included in pa0102. In case they do not, it is not possible to use that table for extraction. You need to create a view on it as well as on the unit tables and include the unit fields into your view. Then it will work.
    kind regards
    Siggi

  • How to extract data from xml and insert into Oracle table

    Hi,
    I have a large xml file. which will have hundreds of the following transaction tags having column names and there values.
    There is a table one of the schema with coulums "actualCostRate","billRate"....etc.
    I need to extract the values of these columns and insert into the table
    <Transaction actualCostRate="0" billRate="0" chargeable="1" clientID="NikuUK" chargeCode="LCOCD1" externalID="L-RESCODE_UK1-PROJ_UK_CNT_GBP-37289-8" importStatus="N" projectID="TESTPROJ" resourceID="admin" transactionDate="2002-02-12" transactionType="L" units="11" taskID="5017601" inputTypeCode="SALES" groupId="123" voucherNumber="ABCVDD" transactionClass="ABCD"/>
    <Transaction actualCostRate="0" billRate="0" chargeable="1" clientID="NikuEU" chargeCode="LCOCD1" externalID="L-RESCODE_US1-PROJ_EU_STD2-37291-4" importStatus="N" projectID="TESTPROJ" resourceID="admin" transactionDate="2002-02-04" transactionType="L" units="4" taskID="5017601" inputTypeCode="SALES" groupId="124" voucherNumber="EEE222" transactionClass="DEFG"/>

    Re: Insert from XML to relational table
    http://www.google.ae/search?hl=ar&q=extract+data+from+xml+and+insert+into+Oracle+table+&btnG=%D8%A8%D8%AD%D8%AB+Google&meta=

  • Extract the data from SQL Server and Import into Oracle

    Hi,
    I would like to run a daily job that will export the table data from SQL server table (it will be only one or two table) and Import back into Oracle table (it might one or two table tables).
    Could you please guide me that how can i do this using either sql server or oracle?
    We have oracle 9.2 and sql server 2005.
    Normally i do from flat file which is generated by source destination nand i dump into oracle using sql*loader but this time I have to directly extract/export the data from MS Sql server and load into Oracle table, mostly it will reload so i might doing any massaging data during the load.
    If you show me the detail approach, it will be really appreciated.
    I have access to Sql server but i don't how to use sql server to do this or using oracle as a daily job even becuase have to schedule the job for this as it will be a daily job.
    Thanks,
    poratips

    Unless you can find an open source ODBC driver for SQL Server that runs on Solaris (and I wouldn't be overly hopeful there) Heterogeneous Services would require that you license something-- a third party ODBC driver, a new Oracle instance, or an Oracle Transparent Gateway.
    As I stated below, you could certainly use SQL Server's ETL tool, DTS. Oracle's ETL tools would require additional licensing since you're just on 9.2. You could also write a small application (Java or otherwise) that connected to both databases and transferred the data. If you're particularly enterprising, you could load the SQL Server Type 4 JDBC driver into Oracle's JVM and write a Java stored procedure that connected to the SQL Server database via JDBC, but that's a pretty convoluted approach.
    Justin

  • Hi, extract data from xml file and insert into another exiting xml file

    i am searching code to extract data from xml file and insert into another exiting xml file by a java program. I understood it is easy to extract data from a xml file, and how ever without creating another xml file. We want to insert the extracted data into another exiting xml file. Suggestions?
    1st xml file which has two lines(text1.xml)
    <?xml version="1.0" encoding="iso-8859-1"?>
    <xs:PrintDataRequest xmlns:xs="http://com.unisys.com/Anid"
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://com.unisys.com/Anid file:ANIDWS.xsd">
    <xs:Person>
    xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
    xsi:schemaLocation="http://com.unisys.com/Anid file:ANIDWS.xsd">
    These two lines has to be inserted in the existing another xml(text 2.xml) file(at line 3 and 4)
    Regards,
    bubbly

    Jadz_Core wrote:
    RandomAccessFile? If you know where you want to insert it.Are you sure about this? If using this, the receiving file would have to have bytes inserted that exactly match the number of bytes replaced. I'm thinking that you'll likely have to stream through the second XML with a SAX parser and copy information (or insert new information) as you stream with an XML writer of some sort.

  • Extract data from xml file and insert into another exiting xml fil

    hello,
    i am searching extract data from xml file and insert into another exiting xml file by a java program. I understood it is easy to extract data from a xml file, and how ever without creating another xml file. We want to insert the extracted data into another exiting xml file. Suggestions?
    Regards,
    Zhuozhi

    If the files are small, you can load the target file in a DOM document, insert the data from the source file and persist the DOM.
    If the files are large, you probably want to use a StAX or SAX.

  • How to get the data from a file and insert into a table

    Good morning,
    NEED TO READ THIS FILE sui_facturacion_alcantarillado_15085_2011_01_76845_00A.csv containing the following information
    NUID,NUMERO_DE_CUENTA_CONTRATO,CÓDIGO_DANE_DEPARTAMENTO,CÓDIGO_DANE_MUNICIPIO,ZONA_IGAC,SECTOR_IGAC,MANZANA_O_VEREDA_IGAC,NÚMERO_DEL_PREDIO_IGAC,CONDICION_DE_PROPIEDAD_DEL_PREDIO_IGAC,DIRECCIÓN_DEL_PREDIO,NÚMERO_DE_FACTURA,FECHA_DE_EXPEDICIÓN_DE_LA_FACTURA,FECHA_DE_INICIO_DEL_PERÍODO_DE_FACTURACIÓN,DIAS_FACTURADOS,CÓDIGO_CLASE_DE_USO,UNIDADES_MULTIUSUARIO_RESIDENCIAL,UNIDADES_MULTIUSUARIO_NO_RESIDENCIAL,HOGAR_COMUNITARIO_O_SUSTITUTO,USUARIO_FACTURADO_CON_AFORO,USUARIO_CUENTA_CON_CARACTERIZACIÓN,CARGO_FIJO,CARGO_POR_VERTIMIENTO_BASICO,CARGO_POR_VERTIMIENTO_COMPLEMENTARIO,CARGO_POR_VERTIMIENTOSUNTUARIO,CMT,VERTIMIENTO_DEL_PERIOD_EN_METROS_CUBICOS,VALOR_FACTURADO_POR_VERTIDO,VALOR_DEL_SUBSIDIO,VALOR_DE_LA_CONTRIBUCIÓN,FACTOR_DE_SUBSIDIO_O_CONTRIBUCIÓN_CARGO_FIJO,FACTOR_DE_SUBSIDIO_O_CONTRIBUCIÓN_VERTIMIENTO,CARGOS_POR_CONEXIÓN,PAGO_ANTICIPADO_DEL_SERVICIO,DÍAS_DE_MORA,VALOR_DE_MORA,INTERESES_POR_MORA,OTROS_COBROS,CAUSAL_DE_REFACTURACIÓN,NUMERO_DE_LA_FACTURA_OBJETO_DE_REFACTURACIÓN,VALOR_TOTAL_FACTURADO,PAGOS_DEL_CLIENTE_DURANTE_EL_PERÍODO_FACTURADO
    242602,242602,76,845,99,99,9999,9999,999,CLL 5 CRA 7 PEATONAL,24911920,12-01-2011,01-12-2010,30,01,,,0,0,0,1,0000000000.00,0000000000.00,0000000000.00,0000000000.00,0000000005,0000002200.00,0000000000,0000000000,0.000,0.000,0000000000.00,0000000000.00,0,0000000000.00,0000000000.00,0000000000.00,0,0,0000002201.00,0000000000.00
    242604,242604,76,845,99,99,9999,9999,999,CRA 4 # 6 - 13,24911846,12-01-2011,01-12-2010,30,01,,,0,0,0,1,0000000000.00,0000000000.00,0000000000.00,0000000000.00,0000000013,0000002200.00,0000000000,0000000000,0.000,0.000,0000000000.00,0000000000.00,0,0000000000.00,0000000000.00,0000000000.00,0,0,0000002201.00,0000004411.00
    242605,242605,76,845,99,99,9999,9999,999,CRA 2 CLLES 3 Y 4,24911509,12-01-2011,01-12-2010,30,01,,,0,0,0,1,0000000000.00,0000000000.00,0000000000.00,0000000000.00,0000000004,0000002200.00,0000000000,0000000000,0.000,0.000,0000000000.00,0000000000.00,0,0000000000.00,0000000000.00,0000000000.00,0,0,0000002201.00,0000002200.00
    this is the function that I have
    <<function_test>>
    DECLARE
    TOTAL_CAR NUMBER;
    POS_1 NUMBER:= 0;
    POS_2 NUMBER:= 0;
    REST NUMBER:= 0;
    ACUM NUMBER:= 0;
    CADEN VARCHAR2(200);
    nom_archivo varchar2(80);
    v1 utl_file.file_type;
    v2 varchar2(2048);
    BEGIN
         nom_archivo := 'sui_facturacion_alcantarillado_15085_2011_01_76845_00A.csv';
         v1:= utl_file.fopen('PUBLIC_ACCESS',nom_archivo,'R',32767);
         utl_file.get_line(v1,v2);
         SELECT LENGTH(v2) INTO TOTAL_CAR FROM DUAL;
         ACUM:=1;
         POS_1:=0;
         WHILE ACUM <= 60
         LOOP
              select instr(v2, ',', 1, ACUM) PRUEBA
              INTO      POS_2
              FROM      DUAL;
              dbms_output.put_line(' TOTAL POSICION 1--> '|| POS_1);
              dbms_output.put_line(' TOTAL POSICION 2--> '|| POS_2);
              dbms_output.put_line(' TOTAL ACUMULADO --> '|| ACUM);
              REST := (POS_2-POS_1)-1;
              SELECT SUBSTR(v2,(POS_1+1),REST) PRUEBA2
                   INTO CADEN
              FROM      DUAL;     
              dbms_output.put_line(' CADENA SELECCIONADA --> '|| CADEN);
              ACUM := ACUM + 1;
              POS_1:= POS_2;
         END LOOP;
         utl_file.fclose(v1);
         dbms_output.put_line(' -->');
         dbms_output.put_line(' TOTAL POSICION 1-->'|| POS_1);
         dbms_output.put_line(' TOTAL POSICION 2-->'|| POS_2);
         dbms_output.put_line(' TOTAL ACUMULADO -->'|| ACUM);
         dbms_output.put_line(' TOTAL DE CARACTERES -->'|| TOTAL_CAR);
         dbms_output.put_line(' ');     
         EXCEPTION
              WHEN NO_DATA_FOUND THEN
                   dbms_output.put_line('NO SE ENCONTRARON MAS CARACTERES');
              WHEN OTHERS THEN
                   dbms_output.put_line('OTRO TIPO DE ERROR ');
                   dbms_output.put_line('CODIGO ERROR '|| SQLCODE ||' '||SQLERRM);
    END;
    Which must be separated by a comma and enter a table I have the following procedure in which only brings me the first line, which need not
    The current role I have just read and extract data from the row number 1, I do not need information.
    I need information for rows 2,3,4. In each row there are 41 fields, which I enter in a table called Dato_archivos.
    how to perform this function? ...
    I appreciate the cooperation and explanation ...
    GOOD DAY ...
    REYNEL SALAZAR MARTINEZ
    COLOMBIA ...

    When you get an error with external tables (or sql*loader) look in the same folder as the data file and you should get a .log file and maybe a .bad file too.
    The log file should indicate the nature of the error it has trying to load the data.
    I've just copied your sample data from your first post to a file on my server and tried it to find that you are not specifying the required format for your dates. The below shows it now working...
    CREATE TABLE tabla_prueba
      (NUID NUMBER,
       NUM_CUENTA_CONTRATO NUMBER,
       COD_DANE_DD NUMBER,
       COD_DANE_MM NUMBER,
       ZONA_IGAC NUMBER,
       SECTOR_IGAC NUMBER,
       MANZANA_VEREDA_IGAC NUMBER,
       NUM_PREDIO_IGAC NUMBER,
       CONDICION_PREDIO_IGAC NUMBER,
       DIRECCION_PREDIO_IGAC VARCHAR2(80),
       NUM_FACTURA NUMBER,
       FECHA_EXPED_FACTURA DATE,
       FECHA_INI_PERIODO_FACTURACION DATE,
       DIAS_FACTURADOS NUMBER,
       COD_CLASE_USO NUMBER,
       UNI_MULTIUSUARIO_RESIDENCIAL NUMBER,
       UNI_MULTIUSUARIO_NORESIDENCIAL NUMBER,
       HOGAR_COMUNITARIO NUMBER,
       USUARIO_FACTURADO_AFORO NUMBER,
       USUARIO_CON_CARACTERIZACION NUMBER,
       CARGO_FIJO NUMBER,
       CARGO_VERTIMENTO_BAS NUMBER,
       CARGO_VERTIMENTO_COMP NUMBER,
       CARGO_VERTIMENTO_SUNT NUMBER,
       CMT NUMBER,
       VLR_FACTURADO_VERTIDO NUMBER,
       VLR_SUBSIDIO NUMBER,
       VLR_CONTRIBUCCION NUMBER,
       FACTOR_SUBS_CONTR_CARGO_FIJO NUMBER,
       FACTOR_SUBS_CONTR_VERTIMENTO NUMBER,
       CARGO_CONEXION NUMBER,
       PAGO_ANTICIPADO_SERVICIO NUMBER,
       DIAS_MORA NUMBER,
       VLR_MORA NUMBER,
       INTERES_MORA NUMBER,
       OTROS_COBROS NUMBER,
       CAUSAL_REFACTURACION NUMBER,
       NUM_FACTURA_OBJ_REFACTURACION NUMBER,
       VLR_TOTAL_FACTURADO NUMBER,
       PAGOS_CLIENTE_DURANTE_PERIODO NUMBER
    ORGANIZATION EXTERNAL
      (TYPE ORACLE_LOADER
       DEFAULT DIRECTORY TEST_DIR
       ACCESS PARAMETERS
        (RECORDS DELIMITED BY NEWLINE
         SKIP 1
         FIELDS TERMINATED BY ","
         OPTIONALLY ENCLOSED BY '"'
          (NUID,
           NUM_CUENTA_CONTRATO,
           COD_DANE_DD,
           COD_DANE_MM,
           ZONA_IGAC,
           SECTOR_IGAC,
           MANZANA_VEREDA_IGAC,
           NUM_PREDIO_IGAC,
           CONDICION_PREDIO_IGAC,
           DIRECCION_PREDIO_IGAC,
           NUM_FACTURA,
           FECHA_EXPED_FACTURA CHAR DATE_FORMAT DATE MASK "DD-MM-YYYY",
           FECHA_INI_PERIODO_FACTURACION CHAR DATE_FORMAT DATE MASK "DD-MM-YYYY",
           DIAS_FACTURADOS,
           COD_CLASE_USO,
           UNI_MULTIUSUARIO_RESIDENCIAL ,
           UNI_MULTIUSUARIO_NORESIDENCIAL,
           HOGAR_COMUNITARIO ,
           USUARIO_FACTURADO_AFORO ,
           USUARIO_CON_CARACTERIZACION ,
           CARGO_FIJO ,
           CARGO_VERTIMENTO_BAS ,
           CARGO_VERTIMENTO_COMP,
           CARGO_VERTIMENTO_SUNT,
           CMT,
           VLR_FACTURADO_VERTIDO,
           VLR_SUBSIDIO ,
           VLR_CONTRIBUCCION ,
           FACTOR_SUBS_CONTR_CARGO_FIJO ,
           FACTOR_SUBS_CONTR_VERTIMENTO ,
           CARGO_CONEXION ,
           PAGO_ANTICIPADO_SERVICIO ,
           DIAS_MORA ,
           VLR_MORA ,
           INTERES_MORA ,
           OTROS_COBROS ,
           CAUSAL_REFACTURACION ,
           NUM_FACTURA_OBJ_REFACTURACION,
           VLR_TOTAL_FACTURADO,
           PAGOS_CLIENTE_DURANTE_PERIODO
        LOCATION ('test.csv')
    SQL> select * from tabla_prueba;
          NUID NUM_CUENTA_CONTRATO COD_DANE_DD COD_DANE_MM  ZONA_IGAC SECTOR_IGAC MANZANA_VEREDA_IGAC NUM_PREDIO_IGAC CONDICION_PREDIO_IGAC DIRECCION_PREDIO_IGAC                                                    NUM_FACTURA FECHA_EXPE FECHA_INI_
    DIAS_FACTURADOS COD_CLASE_USO UNI_MULTIUSUARIO_RESIDENCIAL UNI_MULTIUSUARIO_NORESIDENCIAL HOGAR_COMUNITARIO USUARIO_FACTURADO_AFORO USUARIO_CON_CARACTERIZACION CARGO_FIJO CARGO_VERTIMENTO_BAS CARGO_VERTIMENTO_COMP CARGO_VERTIMENTO_SUNT        CMT
    VLR_FACTURADO_VERTIDO VLR_SUBSIDIO VLR_CONTRIBUCCION FACTOR_SUBS_CONTR_CARGO_FIJO FACTOR_SUBS_CONTR_VERTIMENTO CARGO_CONEXION PAGO_ANTICIPADO_SERVICIO  DIAS_MORA   VLR_MORA INTERES_MORA OTROS_COBROS CAUSAL_REFACTURACION NUM_FACTURA_OBJ_REFACTURACION
    VLR_TOTAL_FACTURADO PAGOS_CLIENTE_DURANTE_PERIODO
        242602              242602          76         845         99          99                9999         9999              999 CLL 5 CRA 7 PEATONAL                                                         24911920 12-01-2011 01-12-2010
                 30             1                                                                             0               0                           0          1                    0                     0                     0          0
                        5         2200                 0                            0                            0              0                        0          0          0            0            0                    0                             0
                      0                          2201
        242604              242604          76         845         99          99                9999         9999              999 CRA 4 # 6 - 13                                                               24911846 12-01-2011 01-12-2010
                 30             1                                                                             0               0                           0          1                    0                     0                     0          0
                       13         2200                 0                            0                            0              0                        0          0          0            0            0                    0                             0
                      0                          2201
        242605              242605          76         845         99          99                9999         9999              999 CRA 2 CLLES 3 Y 4                                                            24911509 12-01-2011 01-12-2010
                 30             1                                                                             0               0                           0          1                    0                     0                     0          0
                        4         2200                 0                            0                            0              0                        0          0          0            0            0                    0                             0
                      0                          2201
    SQL>
    {code}                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                   

  • How can I read the bootstrap files and extract the fragment-URLs and fragment-numbers in plain text?

    How can I read the bootstrap files of any HDS Live stream and extract the fragment-URLs and fragment-numbers in plain text?
    Could it be that it is some kind of compressed format in the bootstrap? Can I uncompress it wirh  f4fpackager.exe? Could not find any download for f4fpackager.exe. I would prefere less code to do so. Is there something in Java of JavaScript, that can extract the fragment-numbers?
    Thank you!

    Doesn't sound too hard to me. Your class User (the convention says to capitalize class names) will have an ArrayList or Vector in it to represent the queue, and a method to store a Packet object into the List. An array or ArrayList or Vector will hold the 10 user objects. You will find the right user object from packet.user_id and call the method.
    Please try to write some code yourself. You won't learn anything from having someone else write it for you. Look at sample code using ArrayList and Vector, there's plenty out there. Post in the forum again if your code turns out not to behave.

  • CO-PA Extraction into BW

    Hi all,
    We are looking to extracting CO-PA data into BW for use in reporting. I understand that the COPA utilises four tables namely:
    CE1XXXX-Actual line items
    CE2XXXX-Plan line items
    CE3XXXX-Summary table
    CE4XXXX-Segment table
    Some of our reports require line item data - and some require summary level data. Our questions:
    - Are there any standard SAP extractors that can be utilised to pull this data? If so, what are they?
    - It is expected to pull summary data on an hourly basis...and pull line item data on a daily basis. Performance wise, what is the best approach: to have two separate extractors to pull from CE1XXXX and CE4XXXX or to use one extractor to pull from CE1XXXX and summarise within a DSO and extract the summary data from this? As mentioned, summary data will be required every hour.
    Thanks

    Hi,
    1. There are no std extractors for CO-PA as it is a generating application. The 4 tables get generated when the CO consultants create and activate the operating concern. As Praveen mentioned, you would have to create the datasource from KEB0.
    2. I guess we cannot control what data to be extracted, and from which table we extract data, using the CO-PA datasource. While creating the datasource we can choose characteristics from various sections like segement level, segment table, chars from line items, etc. If you need to select and have selected any char from the line items section, the extractor always reads the line items table, CE1****. We cannot make it read any other table. May be you could load the COPA line item data into BW and aggregate within BW.
    If none of the chars from the line items section are chosen except for the mandatory record currency char, then we have the possibility that the extractor could read tables other than CE1****. CO consultants could create summarization levels which are aggregated sets of data. These would be automatically read by the extractor if the mentioned condition is met.
    http://help.sap.com/saphelp_470/helpdata/en/7a/4c40794a0111d1894c0000e829fbbd/content.htm
    Please go through the 'How to COPA...' document.
    By default CO-PA works on a 30 minute safety interval. The extractor doesnot extract the records posted in the last halfhour.

  • My 12" powerbook logics card crashed. I have a 17". How do i retrieve my internet bookmarks from the 12" by putting it into Target mode on the 12" and looking into the hard drive on the 12"?

    My 12" powerbook logics card crashed. I have a 17". How do I retrieve my internet bookmarks from the 12" by putting it into Target mode on the 12" and looking into the hard drive on the 12"? Where on the 12" hard drive do I go to search for them?

    If the 12's logic board is trashed, I don't think you can get into FireWire Target Disk Mode. The computer has to start to get to FWTDM. Will it starte and run at all?
    If it won;t start, you will have to go inside and extract the hard drive. Then get this gadget:
    Newer Technology Universal Drive Adapter USB 3.0/2.0 - connects any 2.5", 3.5" or 5.25" drive
    It allows you to connect a bare drive to another computer that has USB. It will act just like an external disk drive. Assuming the 12's drive is not damaged due to the LB failure, you can retrieve files quickly.

  • How to copy exitsitng extracter into another new extracter?

    Hello,
    Please some one can tel me how to copy the existing  extracter into another new extracter.
    I have requirment as below
    Currect me if i am wrong ,need to enhance the existing extracter and add some new fields,and need to include some masterdata filds also to it.
    For master data i am going to do write routines at Transfermation so that no need to include in my exter for those filds to extracter.
    Please some one guide me in this scenario.
    Chrs

    Hi,
    you do not have to copy the extractor, you can enhance it with fields and fill them via enhancement coding --> rsa6.
    If you want to copy it, you need to check what kind of extractor you have. Is it a function module or just a table extractor.
    for fm: copy the function module via se37 and copy the extract structure via se11. Add your enhancements in the copied structures and coding. Create a new datasource with rso2.
    for table extractor: create a view adding the additionally needed fields. create a new ds with rso2. Surely you need to add some coding in the relevant enhancement.
    regards
    Siggi

  • Automating the mass activity FPBW (pscd open items extraction into BW)

    Hello all -
    We are trying to automate the PSCD mass activity for open item extraction into bw.  We have tried using the t-code FPBW (how you would run it manually), and we have also tried using the mass activity scheduler - rfkk_ma_scheduler.  We can't get it so the key date updates each week.  Can anyone provide some guidance on how to get this automated?
    Many thanks!

    hi buddy,
    go to SBIW go to the setting for application specificdatasourse in that go to contact account recivable, click the fileds for extrction add the field the data to see in to dffkopbw table after mass activity
    regards
    nag

  • Unable to select and extract object...

    I am currently attempting to edit a recently converted raster to vector image. I have the file opened and selected after having used the "live trace" function to revise it and save as a vector image. Now I would like to select and extract a portion of the image to work with but I am neither able to select and extract the object of focus from within the general picture. When I attempt to use the pointer selection tool to select the object of focus, the entire image becomes selected with a blue border. If I try to use the magic wand nothing happens. If I try to use the lasso tool I can draw around the object of focus but when I close the lasso it simply disappears leaving nothing selected. I've even tried to "ungroup" the image but option is grayed out (unavailable). Any insights into what I am doing incorrectly? By the way, the live trace function becomes active whenever I click anywhere within the general image.

    RUConscious,
    Expanding the Live Trace should make it accessible and editable just like any other vector artwork where you may Direct Select individual/multiple Anchor Points, Select paths, etc.
    Maybe a closeup of the issue area, with a hint at what exactly you wish to do, would help the helpers, preferably including the relevant part of the Layers panel (which may be moved).
    Apart from that, others may tell you that I am too old, or silly, to apply Live Trace.

  • Custom Document & Custom Doc Legal Control Extraction into SAP BI

    Dear All,
    After all the business content for GTS was activated in GTS and SAP BI, the following configuration steps were taken in GTS to extract this data into BI for
    reporting:
    [1] Integration with Other mySAP.com Components > Data Transfer to the SAP Business Information Warehouse > General Settings > Maintain Control Parameters for Data Transfer [Used ECC as guide]
    [2] Global Trade Services > General Settings > Organizational Structure > Control Settings at FTO Level for SAP NW Business Intelligence (BI) 
    Inserted new records missing when comparing to Assgmt of Feeder System Grouping Org.Units to LLS Org.Units table [/SAPSLL/TCOOGS]
    Turned on BI Active flag for all records  
    [3] Global Trade Services > General Settings > Document Structure > Define Document Types for Application Areas
    For each document type in all the folders, turned on ‘transfer to SAP Netweaver Business Intelligence Active’ flag where it was available to be turned on.
    [4] Global Trade Services > General Settings > Document Structure > Define Item Categories for Application Areas
    For  each item category in all the folders, turned on ‘transfer to SAP  Netweaver Business Intelligence Active’ flag where it was available to be turned on.
    Focusing on custom document and custom document legal control extractors, the data comes from /SAPSLL/CUIT which contains 113K records in development, however, the amount of data extracted into BI is minimal.  So I have a few questions:
    Did I miss any additional configuration swithces to extract into BI?
    SAP Documentation: You have entered a unit of weight in the table for GTS: Control Settings at Foreign Trade Organization Level. This unit of weight must be the common unit of weight that you want to be used in strategic reporting in BI . The values in the fields for gross weight (WEIGR), net weight (WEINE) and unit of weight (WEIDI)
    depend on the unit of weight you maintain at FTO level in the IMG for converting weights for SAP BIDoes each FTO need to be insert with different unit of measure?  ie FTO123 UoM = EA; FTO123 UoM = CS; FTO123 UoM = KG
    Should all the data records from /SAPSLL/CUIT be extracted into Custom Document DSO [0SLL_DS01]? If no, why? If yes, refer to pt 1 and what am I missing?
    SAP Documentation: Only data that belongs to the legal controls for the following legal regulations is transferred to BI:
    German Foreign Trade Laws
    Prohibitions and Restrictions
    Other What is the significance of these legal controls that SAP has decided to only transfer this type into BI? If the client want additional legal controls, would it be possible to extract them into BI?
    If anyone is currently using the std content for BI reporting, any assistance with my questions would be appreciated.
    Sincerely,
    Claire

    Hi Claire,
    Your customizing looks correct. But the amount of data extracted depends on your condition you have defined during extraction(e.g. Creation Date)
    Also in the coding only /SAPSLL/CUIT~ITSTA = space (Standard Item) and A (Item is allocated) is considered in 0SLL_CD_1.
    You could find out reason by debugging to find out exact reason.
    T-code: RSA2 , Enter Data source : 0SLL_CD_1. Then click display.
    Click Extract check ( Shift + F8 ) and put break point in coding FM : /SAPSLL/BW_GET_CD_1
    Best regards,
    Vincent

  • Need help for SRM-6 Data Extraction into Bi-7

    Hi Experts,
    I am looking for help regarding SRM DataSource Extraction to BW. I am working on BW project and need to extract data from SRM Data Source 0BBP_TD_CONTR_2 and 0SRM_TD_PRO. I need to know about the extraction process from SRM. Is there a tool available like LO Cockpit in SRM for extraction and how can we manage the delta in it. Or i can use the T-code RSA5 for that and follow the normal process.  If I can get some documentation that can give me an idea for data extraction and delta management  in SRM that would be appreciated. Any kind of help will be welcome.
    Thanks in Advance.

    Hi,
    May help you:----
    How to extract data from SRM into BW system.
    SRM Audit Management Data Extraction into BW
    Procedure for standard data source extraction from SRM to BW 3.5.
    Regards,
    Suman

Maybe you are looking for