Copy DTP

Hi All,
Can we copy the DTP. I have got some filters in my DTP and the same need to be apllied to the new DTPs.
Please lt me know if we con copy and the procedure.
Thanks
Indira.S
Message was edited by:
        Indira SR

Hi,
Just happened to see this thread (though closed).
Copy DTP is possible with a bit of coding.
Check this How to paper:
[Create Copy DTP|https://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/30cbe0f2-e8c1-2b10-e292-bd30bc3ea672]
Paper by : Shabar Alibhai
P.S : The thread is old but this paper was somewhere in 2008 end.
Edited by: Praveen G on Mar 16, 2009 9:47 AM

Similar Messages

  • After system copy DTP pointing to Prod Sourcesystem in Quality

    Hi,
    We have copied BWP system to BWQ.
    All the transformations are correctly pointing to R3Q system.
    But DTPs are still pointing to R3P system.
    How to convert this source system assignment for all the DTPs and InfoPackages?
    Regards
    SSS

    Hi,
    Collect the DTPs and then Transport the DTPs again.
    See the following  SAP Notes:
    SAP Note 886102 - System Landscape Copy for SAP NetWeaver BW
    SAP Note 325470 - Activities after client copy in BW source systems
    SAP Note 325525 - Copying and renaming systems in a BW environment
    SAP Note 1333302 - Special procedure for BW system copy
    SAP Note 996238 - Dump 'RFC connection error' after BW system copy
    Search for PDF in SDN
    How Tou2026System Copy in SAP Business Intelligence System Landscapes
    Thanks
    Reddy

  • Urgent Issue - Copy DTP and Process Chains in New Client Instance

    We have created new client Instance ( CLNT 510 ) based on our present client ( CLNT 500 ).
    We have all the process changes, application components from source system CLNT 500 and data ( both for master and transaction data ) in place for the CLNT 500.
    Now we need to use the CLNT 510 for our futher development. RFC connection has been established.
    Just want to know, do we need to create all the Application component trasfer , Transformation, DTP , Infopackages , Process chains again for the CLNT 510 data.
    please help.
    Thanks and Kind Regards,
    Pratap Gade.

    Hello Jr Roberto,
    Hello Jr Roberto,
    We have company implementing mutiple SAP Implementations in different countries.
    Client for BW Dev is 400.
    Source system for Client BWD :
    ECC Client ERP Dev - 500 ( for Holland Implementation )
    ECC Client EPR Dev - 510 ( for German Implementation  - Instance of ECC 500)
    We are completed with the Client 500 implementation.
    We are starting the german implementation and they have created a intance of Holland ECC Dev client 500 for german implementation with german data.
    Now they have setup 2 rfc connection in BWD 400.
    Also finally when they move to QAS client they are going to integrate the data of the both countries into one client.
    Presenlty I need to use ECC client 510 client for our german development.
    We have around 40 SAP BI reports in our first implementation. and around 10 in our second implementation.
    Just want to know, what are issues to be considered in loading master data ( will it overwrite the master data from the client 500, if i extract the master data from client 510 )...and all the other issues like can I copy all the process chains...etc  from client ECC 500 to ECC 510, without creating all transformation, DTP , Infopackages to load all the master from client 510 again.
    I need a broad understand on how to go about the development ...starting with Application Components transfer from source system Client 510 ....to report development. ( if possible using the already created objects....infopackages , DTP etc..)
    your help would be really appricated Mr Roberto.
    Thanks and Kind Regards,
    Shekar.

  • Error while copying a DTP

    Hi Experts,
    Its a simple yet confusing error to me. I copy a DTP which has some filters defined. I get an error saying both the DTPs can not have the same filter criteria - if its a Delta DTP. I say ok and try to change the filter values , interestingly the error still remains and my new DTP does not activate.
    Please suggest how I can resolve this. I can not change the udpate method to "Full". It has to be a "Delta".
    Appreciate your inputs,
    thanks,
    Shweta

    Hi Shweta,
    I think you have posted your question to the wrong forum.  The OnDemand forum is specifically for applications like crystalreports.com and bi.ondemand.com.
    Cheers
    Steve

  • BW Procedure System Copy - Problemas with transformations and DTP

    Hello,
    My Administration system have done a BW production copy to a BW non-productive system. After that, I have created a transport order in BW production with the option "with the source-system-dependent objects". Before I transport this order to BW non-productive system, I have deleted, in this system, the ancient source-system assignment.
    When I have transported the order to BW non-productive system, the system I created transformation an DTP with another  ID. I was expected that the ID was the same and not another one. Is this normal? What will happen when i tranport the changes form BW non-productive system to BW productive system?
    Best regards.
    Sérgio Pinto

    Hello Raghavendra,
    But the BDLS report is not only for changing the logical name of the BW system? Is it possible to use this program in BW for changing the logical name of the source system than is assigned to the datasources?
    By changing ID I want to say that the DTP with the ID DTP_4EC3YB6B6DT0OIH1TMBB0F5VG is change to the ID DTP_4K1H1XZ38M1CLVJ2IA8KFDQYY in BW no-productive system, after transport from productive to non-productive system.
    Best regards.
    Sérgio Pinto

  • DTP are inactive after System copy

    Hi Guys!!!
    I have done a system copy from one sbox another. After that all the DTP are inactive after System copy.
    Please help me in this regard.
    Thanks in advance
    Peter..

    Hi,
    If u copy the transformation get inactive, bcoz it has changed from existing target to new target.
    so need to activate all.
    Hope it helps u,
    Thanks

  • When we do system copy how the DTP converted

    Hello All,
    I am little curious to know about the DTP bahaviour after the system copy.
    example: In DEv system I created the DTP A
    When it went to Quality it become B
    When it went to Productiion it become C
    This is normal behaviour in 7.0.
    When we do the system copy the DTP C will come in DEV and Quality system.
    Next time suppose i changed the DTP C and I will transport it to Quality does it create new DTP in Quality or it overwrites in quality.
    When we run the BDLS aftere sytsem copy will DTP technical name will change back to A in developement?
    I am little curious to know about this.
    Any help will be appreciated.
    Thanks
    Gopal

    Hi Chetan,
    I already checked the note but it has not mentioned the behaviou of DTP after suystem copy. It is giving the logical name with DTP and I ma talking of the DTP Technical Name.
    I am really curious to know about this.
    If any one did the system copy and can some one throw the light on tha. Really appreciate.
    Thanks
    Gopal

  • Is there a way to copy a DTP?

    Is there a way to copy a DTP so that I have several that are exactly the same except for the description?
    Thnaks

    Hi
    Only one DTP is there betweena  source and a target. You can have a number of transformations and infosources if required, but DTP can only be 1.
    So, you cant create a copy of DTP
    Regards
    Shilpa

  • Copy transformations and DTPs

    Is there a way to copy one Transformation to another? Is there also a way to copy one DTP to another?

    Transformations can be copied, but DTPs cannot be copied. To copy a Transformation, do the following:
    1) Right-click on the Transformation that is to be copied
    2) Select the Object Type and enter the Name of the source and target for the new Transformation
    3) Hit Enter
    4) Make any necessary updates to the mappings and transformational logic, as needed

  • Can I link DTP version content to mobile version without copy / paste it?

    I am creating site and I am little bit frustrated to copy / paste text between desktop and mobile version.
    Is it bossible, that text content are linked together in both versions? It would be great if  all text content (not formatting) changes updates automatically to both versions?
    Thanks again.
    T: Ari

    Hi,
    Please take a look to the similar discussion
    Muse, Desktop copy to tablet, Contents are not displayed

  • Data not loading to cube copied from 0sd cubes

    hello everyone,
    wen ever im trying to load data from r/3 into a cube which is created by copying from a 0sd cube, everything is running successfully, info package, dtp, transformations and all, but when i look into the contents tab, it says this is not a basis cube, can anyone explain pls.
    thanks in advance

    Hi
    Some one changed the settings of your cube from Basic cube to transactional cube ( real time infocube in IP)
    Change it from transactional cube back to basic cube..This error will not come
    Regards
    N Ganesh

  • ASSERTION_FAILED when Activate a DTP

    I got an error message when trying to activate a DTP. Does anyone know how to fix it? Thanks!
    Runtime Errors         ASSERTION_FAILED
    Date and Time          03/27/2007 14:29:57
    Short dump has not been completely stored (too big)
    Short text
         The ASSERT condition was violated.
    What happened?
         In the running application program, the ASSERT statement recognized a
         situation that should not have occurred.
         The runtime error was triggered for one of these reasons:
         - For the checkpoint group specified with the ASSERT statement, the
           activation mode is set to "abort".
         - Via a system variant, the activation mode is globally set to "abort"
           for checkpoint groups in this system.
         - The activation mode is set to "abort" on program level.
         - The ASSERT statement is not assigned to any checkpoint group.
    Error analysis
         The following checkpoint group was used: "No checkpoint group specified"
         If in the ASSERT statement the addition FIELDS was used, you can find
         the content of the first 8 specified fields in the following overview:
         " (not used) "
         " (not used) "
         " (not used) "
         " (not used) "
         " (not used) "
         " (not used) "
         " (not used) "
         " (not used) "
    Trigger Location of Runtime Error
         Program                                 CL_RSAR_PSA===================CP
         Include                                 CL_RSAR_PSA===================CM006
         Row                                     152
         Module type                             (METHOD)
         Module Name                             UPDATEDIRECTORY_TABLES
    Source Code Extract
    Line  SourceCde
    122               i_uni_idc25       = l_codeid
    123               i_program_class   = 'RSAR_ODS_MAINTAIN'
    124             EXCEPTIONS
    125               deletion_rejected = 2
    126               OTHERS            = 3.
    127         ENDIF.
    128       ENDIF.
    129       UPDATE rstsods SET tstpnm   = sy-uname
    130                 timestmp  = l_s_ods-timestmp
    131                 userapp   = p_userapp
    132                 userobj   = p_userobj
    133                 maintprog = ''
    134            WHERE odsname = l_s_odsfield-odsname
    135            AND   version = l_s_odsfield-version.
    136       IF sy-subrc = 0.
    137         IF i_partitioned = rs_c_true.
    138 *--   Entry could exist but includes no partition number,
    139 *     because the PSA was not partitioned before
    140           l_tablnm = p_psa_techname.
    141
    142           CALL FUNCTION 'RSDU_PARTITIONS_INFO_GET'
    143             EXPORTING
    144               i_tablnm              = l_tablnm
    145             IMPORTING
    146               e_ts_part_info        = l_ts_part_info
    147             EXCEPTIONS
    148               table_not_exists      = 1
    149               table_not_partitioned = 2
    150               OTHERS                = 3.
    151
    >>>           ASSERT sy-subrc = 0.
    153
    154           DESCRIBE TABLE l_ts_part_info LINES l_num_partitions.
    155           READ TABLE l_ts_part_info INDEX l_num_partitions INTO l_s_part_info.
    156           l_highest_partvalue = l_s_part_info-high_value.
    157
    158           UPDATE rstsods SET partno  = l_highest_partvalue
    159             WHERE odsname = l_s_odsfield-odsname
    160             AND   version = l_s_odsfield-version.
    161
    162         ENDIF.
    163       ELSE.
    164 *       create new version
    165         l_s_ods-odsname       = l_s_odsfield-odsname.
    166         l_s_ods-version       = i_next_version.
    167         l_s_ods-dateto        = rsods_c_dateto_01019999.
    168         l_s_ods-datefrom      = rsods_c_datefrom_01011998.
    169         l_s_ods-objstat       = rs_c_objstat-active.
    170         l_s_ods-odsname_tech  = p_psa_techname.
    171         l_s_ods-progname      = i_progname.

    I guess it is Notes 1012607.
    Summary
    Symptom
    Note: This note is relevant only for 'ORACLE' and 'MSSQL' database systems. After you implement this note, you must also carry out some manual corrections (see 'Solution', below).
    If you are working with database system DB2 or MSSQL, also implement Note 1022026.
    When data is written or activated or when a DataStore object is activated, the following errors occur:
    Similar errors may also occur for the DataSource and data transfer process (DTP).
    ORA-01502: index 'SAPDAT./BIC/A*KE' or partition of such index is in unusable state
    Column 'PARTNO' is partitioning column of the index '/BIC/A*KE'. Partition columns for a unique index must be a subset of the index key.
    error #RSDU_TABLE_TRUNC_PARTITION_MSS: Error While Calling Module MSS_TRUNC_PARTITION_FROM_TABLE Message no. 0U534#.
    <b>ASSERTION_FAILED in class 'CL_RSAR_PSA'.</b>
    Error message D0 313 in the activation log. The message does not contain any text. In the activation log it is displayed as an empty line with a red traffic light.
    Other terms
    DBIF_RSQL_SQL_ERROR, D0 313, D0313
    Reason and Prerequisites
    Reason:
    The partitioning logic of the persistent staging area (PSA) service does not recognize that the PARTNO field must not be deleted.
    For write-optimized DataStore objects, the active table is created as a partitioned table, even though a global index is used to ensure uniqueness of data. This is not compatible with the 'drop of a partition'.
    In the DataSource maintenance, you have the option to define key fields. For the first 16 key fields of the DataSource field list, a global index is also created.
    If 'semantic groups' are used in the DTP, the error stack is created with a global index.
    Solution
    Implement the corrections by importing the Support Package or by implementing the advance correction. As a result, the 'range' partitioning is deactivated in the PSA service as soon as a global index is requested.
    The error can occur for the objects: DataStore (only the write-optimized type), DataSource, and error stack of the DTP.
    This note contains the 'RSAR_PSA_PARTITION_CHECK' program, which you can use to analyze the objects. Execute the program. Use the search strings listed in section 5), depending on whether you want to analyze individual objects or object types. If you do not make an entry in the PODSTECH field (technical name of the PSA), the system checks all existing PSA tables, which may take some time.
    You can use transaction SLG1 to display the log for 'RSAR_PSA_PARTITION_CHECK'. Select the following:
               Object        = 'RSAR'
    Subobject   = 'METADATA'
    Ext. Identif. = 'RSAR_PSA_PARTITION_CHECK'
    You must make different manual changes to repair each of the different object classes.
    1) DataStore (write-optimized)
    Incorrect DataStores are identified in the log of the check program with the PSA type 'FASTSTORE'. The name after 'Obj:' is the technical name of the corresponding DataStore object.
    For a DataStore object of the 'write-optimized' type, a global index with relation to the semantic key is created if the 'Do Not Check Uniqueness of Data' indicator is not set.
    Check if you need to ensure that data is unique in your scenario.
    1. If you do not need the data to be unique:
                        Set the flag: 'Do Not Check Uniqueness of Data', and activate the DataStore object. The DataStore object is now consistent again.
    2. If you need unique data:
    In this case, you must departition and convert the table.
    If the error occurred when you activate the DataStore itself or when you activate the data, you must activate the DataStore object after converting the active table. You need the technical name of the active table for the conversion. You can get this directly from the log of 'RSAR_PSA_PARTITION_CHECK'. If you know which DataStore contains errors, find the technical name of the active table in the Maintain DataStore screen by choosing:
               <Extras> ->
              'Information (logs/status)
    Choose 'Dictionary DB status' to access the status POPUP. You can find the technical name in the 'Active table' field.
    If the table does not contain any data according to the 'RSAR_PSA_PARTITION_CHECK' log, the table is automatically departitioned when you activate the DataStore.
    If the table contains data, you must departition and convert the table as described in section 4.
    After that, use the AdminWorkBench (transaction RSA1) to activate the DataStore object.
    2) DataSource:
    Incorrect DataSources are identified in the log of the check program with the PSA type 'NEW_DS'. The 'Obj:' indicator  is followed by two additional character strings. The first is the technical name of the relevant DataSource. The second is the technical name of the source system.
    PSA tables for DataSources with a key definition must be departitioned.
    The name of the PSA table for the DataSource is contained directly in the 'RSAR_PSA_PARTITION_CHECK' log.
    If the table does not contain any data according to the 'RSAR_PSA_PARTITION_CHECK' log, the table is automatically departitioned when you activate the DataStore.
    If the table contains data, you must departition and convert the table as described in section 4.
    Call transaction 'RSDS' and enter the technical name of the DataSource and the source system and activate the DataSource.
    3) Error stacks for the DTP:
    Incorrect Error Stacks are identified in the log of the check program with the PSA type 'ERRORSTACK'. The 'Obj:' indicator  is  followed by the technical name of the relevant DTP. There may be more than one error stack table for each DTP.
    PSA tables for ErrorStack with a key definition must be departitioned.
    The name(s) of the PSA Error Stack table(s) for the DTP is/are contained directly in the 'RSAR_PSA_PARTITION_CHECK' log.
    If the table does not contain any data according to the 'RSAR_PSA_PARTITION_CHECK' log, the table is automatically departitioned when you activate the DTP.
    If the table contains data, you must departition and convert the table as described in section 4.
    Now call transaction RSDTP, enter the technical name of the DTP and activate the DTP.
    4) Departitioning and converting
    The following manual conversion using transaction SE14 is supported only for ORACLE database systems. Open a problem message under component BW-SYS-DB-MSS if you need to convert tables on a MSSQL database system.
    Call transaction SE14 (Database Utility) for the tables you need to convert. Select 'Table', enter the technical name of the table and choose 'Edit'.
    On the next screen, choose 'Storage Parameters' (Shift+F6).
    On the next screen (Storage Parameters), choose 'For new creation' (F8).
    In the dialog box that then appears, select 'Current database parameters' and copy it by choosing 'Enter'.
    You now get an overview of the storage parameters <Tables>, <Indexes> and existing <Partitions>.
    Under the 'Table' node, if the content of the 'TABLESPACE' field is initial, enter the value from the 'TABLESPACE' field of the first partition.
    For the field 'PARTITIONED BY', choose the option 'No partitioning' and save your changes.
    Exit the screen with the storage parameters.
    On the next screen, ensure that the 'Save data' radio button after 'Activate and adjust database' is selected, then execute the conversion. You execute the conversion by choosing 'Force Conversion' in the <Extras> menu.
    Next, you must correct the PARTNO indicator in table RSTSODS. To do this, call transaction RSRV and execute the test 'Consistency Between PSA Partitions and SAP Administration Information'. You can find this test in transaction RSRV under
    <All Elementary Tests>
                 -> <PSA Tables>
    You can execute the RSRV test and repair for all converted tables at once. For further information about how to use transaction RSRV in this case, see the online documentation. You can call the online documentation by choosing the 'Info' icon.
    5) Search strings:
    a) Use the search string '/BI+/B*' to find the relevant entries for the DataSource, the change logs and the error stack.
    b) Use the search string '/BI+/A*00' to find the relevant entries in the active tables for the DataSource objects.
    SAP NetWeaver 2004s BI
               Import Support Package 13 for SAP NetWeaver 2004s BI (BI Patch 13 or SAPKW70013) into your BI system. The Support Package is available once Note 991093 "SAPBINews BI 7.0 Support Package 13", which describes this Support Package in more detail, has been released for customers.
    In urgent cases, you can implement the correction instructions as an advance correction.
    You must first implement Notes 932065, 935140, 948389, 964580, 969846, 975510, 983212 and 1000448, which provide information about transaction SNOTE. Otherwise, problems and syntax errors may occur when you deimplement certain notes.
    To provide information in advance, the notes mentioned above may already be available before the Support Package is released. In this case, the short text of the note still contains the words "Preliminary version".
    Before you implement an advance correction (if one exists and you want to implement it), see Note 875986. This contains notes regarding the SAP Note Assistant and these notes prevent problems during the implementation.

  • Short Dump "ITAB_DUPLICATE_KEY" while executing DTP

    Hi all,
    i am getting a short dump when i try to execute the DTP to Cube, following is the error details.
    i could not analyze the reason, need your inputs on this.
    Runtime Errors         ITAB_DUPLICATE_KEY                                           
    Date and Time          02/08/2009 21:33:53                                          
    Short dump has not been completely stored (too big)                                                                               
    Short text                                                                               
    A row with the same key already exists.                                                                               
    What happened?                                                                     
         Error in the ABAP Application Program                                                                               
    The current ABAP program "SAPLRSAODS" had to be terminated because it has      
         come across a statement that unfortunately cannot be executed.                                                                               
    Error analysis                                                                     
         An entry was to be entered into the table                                      
          "\FUNCTION=RSAR_ODS_GET\DATA=L_TH_ISOSMAP" (which should have                 
         had a unique table key (UNIQUE KEY)).                                          
         However, there already existed a line with an identical key.                                                                               
    The insert-operation could have ocurred as a result of an INSERT- or           
         MOVE command, or in conjunction with a SELECT ... INTO.                                                                               
    The statement "INSERT INITIAL LINE ..." cannot be used to insert several       
          initial lines into a table with a unique key.                                                                               
    thanks,
    Rk

    Hello Rk,
    What is the data source??
    whats is your SP level??
    Seems like the data source is not properly activated or duplicate entries exists in the system tables for the same data source...
    If you have recently done a system copy??
    Try to activate the data source again...if its from R/3 then replicate and activate it again...also if its a BW object then try to reactivate it as well.
    Thanks
    Ajeet

  • Abap Routine in DTP Filter with selection in a table

    Hi guys,
    I need help please.
    I'm trying include a abap routine in a DTP filter, for this case I need to make a select in a dso table and return a list of criterias.
    Example: for this characteristic 0GL_ACCOUNT i need in a fiter 20 or more 0GL_ACCOUNT of table  "/BIC/DSO_XXX".
    How can I select more than one 0GL_ACCOUNT in a tranparency table  "/BIC/DSO_XXX"... and put in a DTP Fiter.
    DTP FILTER ROUTINE.
    data: l_idx like sy-tabix.
              read table l_t_range with key
                   fieldname = 'GL ACCOUNT'.
              l_idx = sy-tabix.
              if l_idx <> 0.
                modify l_t_range index l_idx.
              else.
                append l_t_range.
              endif.
              p_subrc = 0.

    Try this:
    DATA: lv_rows TYPE n LENGTH 10,
            it_zbw_pl_proj LIKE STANDARD TABLE OF /BIC/DSO_XXX 
            lw_zbw_pl_proj LIKE LINE OF it_zbw_pl_proj .
      SELECT COUNT(*) INTO lv_rows FROM  /BIC/DSO_XXX Where <your condition> .
    SELECT * INTO TABLE it_zbw_pl_proj FROM zbw_pl_proj where <your condition>
      IF lv_rows <> 0.
        LOOP AT it_zbw_pl_proj INTO lw_zbw_pl_proj  .
          l_t_range-iobjnm = '/BI0/GL_ACCOUNT'.
          l_t_range-fieldname = 'GL_ACCOUNT'.
          l_t_range-sign = 'I'.
          l_t_range-option = 'BT'.
          l_t_range-low = lw_zbw_pl_proj-GL_ACCOUNT.
          l_t_range-high = lw_zbw_pl_proj-GL_ACCOUNT.
          APPEND l_t_range.
        ENDLOOP.
      ELSE.
        " No data found for Current Forecast Version.
        p_subrc = 4.
      ENDIF.
    I just copied this data from a DTP routine where i am fetching Versions from a DB table zbw_pl_proj. You may need to change the naming convention and performance tune the code ( like defining internal table only to hold GL_account and then only selecting GL_Account from ODS).
    Hope this helps!
    Regards
    Amandeep Sharma
    Edited by: AmanSharma123 on Jul 14, 2011 2:42 PM

  • Inventory data load from Inventory Cube to another Copy Cube

    Hello Experts,
    I am trying to load Inventory data from the Inventory cube(say YCINV) to a copy cube  say YCOPY_CINV(copy of Inventory cube), but the results appear inconsistant when I compare the reports on these 2 cubes. I am trying to populate a fiield in copy cube so that I can populate the same data back to the original cube with the new field in it, I am doing this reload back and forth for historical data purpose only.
    I have seen lot of posts as how to run the setups for Inventory data, but my case does not need to perform set up runs.
    Does the note 1426533 solve the issue of loading from one cube to another ? we are on SAP BI 7.01 with SP 06 ,but the note specifies SP 07 ?
    I have tried note 375098 to see if it works, but I do not see the options as mentioned from  step "Using DTP (BW 7.x)" in BI to perform this note
    Please advise on whether to go with implementing note 1426533 or is there any other way to load inventory data from one cube to other.
    Regards,
    JB

    Hi Luis,
    Thanks for your reply,
    I do not see any setting like Intial stock in DTP except "initial non-cumulative for non- cumulative". I did try using the option "initial non-cumulative for non- cumulative" ,but the results still do not match with the inventory cube data.I do not see the check box for marker in the copy cube (under roll up tab). Please let me know if we can really implement this solution ,i.e. copying from inventory cube to copy cube and then re-loading it back to inventory cube for historical data. Currenlty, I am comparing the queries on these 2 cubes, if the data matches then I can go  ahead and implement it in Production, other wise it would not be wise to do so .
    Regards,
    JB

Maybe you are looking for

  • Excise Invoice Number  Series Change

    Hi Gurus Its very close to end of financial year on 31st march and according to the rules of Excise.. the excise number series should be changed .. so plz anyone suggest me how this number series is maintained and how it can be changed

  • Will the fire wire western digital drive work on a new imac

    I am now using 3 3 WD My Book Studio FireWire External Hard Drives on my older iMac. I am thinking of getting a new iMac and see that they do not have a firewire plug. Is there a way to use these drives with a new iMac?

  • How to send multiple data files / music files using bluetooth / whatsapp / emails...

    hello everyone,,, is there any option to send multiple data files / music files using bluetooth / whatsapp / emails... as option to select multiple files using "SELECT in menu option / left aA"+ scroll tracepad" is availble with pictures only. and wh

  • Upload BLOB through Handheld application

    Hi, I am developing NW Mobile Handheld application with 7.10 SP10 IDE. I want to upload an image and send it to DOE. I have created a model in DOE and imported the same in NWDS. The data type for the attachment attribute generated is BLOB as expected

  • Lossless to lossy and back?

    I have recently converted my itunes collection (that I burned from CD) to Apple Lossless. My one ipod is big enough to carry the volume and it makes for better listening when I run it thru my system. If I want to put the same tunes on another ipod fo