Issues with Export/Import using Database & Shared buffer

Hi All,
I have a method that calls a program via a job and I am having issues passing data between the two.
Note that two different users process the method and program (via job) resp. This is how I am calling the second prog-
SUBMIT ZPROGRAM
             VIA JOB     l_jobname
                 NUMBER  l_jobcount
                 USER    i_user
             AND RETURN.
I need to pass data from method to  the second program and vice versa and then the method continues its processing with the data acquired from the second program.
I have tried using Import/Export using database and also shared buffer. What I have found is that most of the times I am able to pass data from method to program. However the job takes a couple of min to execute and I think that the data is not making back to the method in time.
I have looked at some useful forum links-
Problem with export/import in back ground
Re: EXPORT/IMPORT  to MEMORY
but havent been able to find an answer yet. Any solution? Thanks in advance for your help!
Liz

Hi Suhas, Subhankar
I have tested the scenario without the job previously itself and it works. Thats the reason, i am trying with the job now as my requirement is that I need to change the user while executing the second report.
Here is an example of my import/export - I am passing the return value from the second report to the first.
Code in second report-
DATA: INDXKEY LIKE INDX-SRTFD VALUE 'RET1'.
INDX-AEDAT = SY-DATUM.
INDX-USERA = SY-UNAME.
EXPORT RETURN1 TO SHARED BUFFER INDX(ST) ID INDXKEY.
Code in first report -
SUBMIT ZPROGRAM
                 VIA JOB     l_jobname
                     NUMBER  l_jobcount
                     USER    i_user
                 AND RETURN.
Once Job close FM is executed successfully, I import the values as follows
IMPORT RETURN1 TO RETURN1 FROM SHARED BUFFER INDX(ST) ID INDXKEY3.
INDXKEY is having value RET1.
However Return1 is not having any values in first report. It has some value in executed without the job
Please note that I have tried Export/import with Database too and I am getting the same results.
Thanks for your suggestions.
Regards, Liz

Similar Messages

  • EXPORT / IMPORT  TO/FROM SHARED BUFFER

    Hello all,
    I am facing a problem with the EXPORT/IMPORT to SHARED BUFFER statements.
    In my report program , I export data to the shared memory.
    I then call a transaction to park an accouting document.
    The BTE 2218 gets triggered in the process. Here the IMPORT works fine.
    Later, there is a standard function module which is called IN UPDATE TASK.
    Within this, the IMPORT statement fails.
    It works on one server but not on another.
    Notes :
    The IMPORT works in debugging mode but fails if I simply run.
    Another point is that the ID used for identifying the shared memory uses sy-uname.
    Can the visiblity of sy-uname in UPDATE TASK be controlled by settings ?
    Any ideas on this ?
    Please don't copy paste the help on SHARED BUFFER etc.
    Thanks in advance.

    Hi Mariano,
    the issue is to due to multiple servers present where SHARED MEMORY is specific for each application server.
    So we export data into shared memory in program A, we have to be sure, that program B or FM which is called in background or update task by program A runs on the same application server
    Here, the problem is when program A calls the program B or FM in background or update it’s a dynamic scheduling to all application server with have batch work processes and not the same application server that of calling program A always, so program B runs on another application server which has different shared  memory.
    Solution will be:-
    To Force program B to run on same application server as of calling program A by
    passing sy-host of calling program A to Function module “JOB_CLOSE” parameter
    name “TARGETSERVER”. OR
    Instead of using SHARED MEMORY we will use DATABASE.
            EXPORT itab FROM itab  TO DATABASE indx(ar) CLIENT sy-mandt ID job_number in programA where job number is unique.
            Then IMPORT itab TO itab FROM database indx(ar) CLIENT sy-mandt ID job_number  in program B Where job number is passed from program A to B.
            Then DELETE FROM DATABASE indx(ar) CLIENT sy-mandt ID job_number.
    Regards,
    Vignesh Yeram

  • Issue with Export / Import Configuration Data in AC 5.3 migration to AC 10

    Hi,
    We are migrating from AC 5.3 to AC 10. As per the Migration Guide I have exported the data from AC 5.3.
    The export created around 57 files on the destination with .dat extension.
    Moved these files to target server Import location.
    Then I have used the utility (GRAC_DATA_MIGRATION) in AC 10. Selected the Import location. and clicked Get Files.
    It has uploaded some files (21 out of 57).
    When I analyzed the uploaded files, I found it has uploaded only "Common Configuration" data. For the below Objects, though the export utility has created files, Import utility did not uploaded the files.
    1. Business Unit
    2. Critical Roles and Profiles
    3. CUP Role Repository
    4. Mitigation Controls
    5. SOD Rules
    6. Workflow
    Does anyone have any idea about this issue?
    Regards,
    Deepak

    Deepak,
    Can you share with me how you are importing the config data? How you are using the path? Please share this and also try toshare how you systems are distributed.
    Regards,
    Faisal

  • Issues with Exporting files using Lightroom 5

    I bought Lightroom 5 for Mac a couple of months ago and have been using it without any issues, then out of the blue I ran into export problems. Sometimes the exported images wouldn’t have the sharpness adjustments, sometimes it wouldn’t have the noise adjustments, sometimes the saturation or vibrancy, and then sometimes a combination of the above. I know there’s a known issue that effects noise on export if you export on files sizes less than 1/3 of the original size, but mine have always been exported at full size with image quality 70, exporting as jpeg.
    I installed the latest update and that didn’t help. I uninstalled the software and that didn’t help either. I then uninstalled everything completely (including the library) by searching both lightroom and adobe in finder > system files> are included and deleted everything I could find pertaining to adobe lightroom (I didn’t delete everything adobe as I have other adobe products). This didn’t help either. What is really bizarre is that I tried restoring my Mac back to a time when Lightroom worked fine, and this didn't solve the issue, and I've also tried wiping the hard drive and just installing lightroom and this didn't work. This really baffled me.
    What I have also found with Lightroom (and I'm sure it's all somehow linked) is that if I'm working on a file, and either close LR down and re-open it, or go to another file and then back to the one I was working on, initially the image has lost it's sharpening again, but if I click on the sharpen slider (without moving it) the sharpening comes back.
    So has anyone any ideas on what my issues are and how to fix them. I’ve tried the advice on here too [url]http://helpx.adobe.com/lightroom/kb/manually-remove-lightroom.html[/url].
    My workaround at the minute is to use full screen view when I've completed my adjustments and take a hi-res screen shot, but the problem is that the saved images are only approx 1600 x 1000 pixels, which is fine for viewing on my Macbook pro but not good if I want to zoome etc.

    snerkler wrote:
    As I said in my first paragraph I don't think it's the known issue...
    Sorry snerkler - I didn't read closely enough .
    Sounds like you have a different problem - if it's a bug, be sure to report it:
    http://feedback.photoshop.com/photoshop_family/products/photoshop_family_photoshop_lightro om
    Rob

  • Hi.  I am having issues with copying files to my shared WB 2TB HDD connected to my airport extreme.  Comes up with error 50.  I am using a Macbook Pro to copy from so not sure what I am doing wrong.  Can someone help? thanks Rory

    Hi.  I am having issues with copying files to my shared WB 2TB HDD connected to my airport extreme.  Comes up with error 50.  I am using a Macbook Pro to copy from so not sure what I am doing wrong.  Can someone help? thanks Rory

    These links might provide some information that may be of help.
    http://support.apple.com/kb/TA20831
    https://discussions.apple.com/message/2035035?messageID=2035035
    I've encountered this error myself upon occasion.  If I remember correctly, it was a permissions/ownership issue with the some of the files I was copying.

  • Problem with export/import in back ground

    Hi Experts,
    I am having a requirement in which i am exporting an internal table to memory from one program and importing that in another program. in foreground it is working fine. But in background it is not working. Is there any work around.
    Here is sample code.
    Program1:
    types:
      begin of tab_type,
        para type string,
        dobj type string,
      end of tab_type.
    data:
      id    type c length 10 value 'TEXTS',
      text1 type string value `IKE`,
      text2 type string value `TINA`,
      line  type tab_type,
      itab  type standard table of tab_type,
        itab1  type standard table of tab_type.
    line-para = 'P1'.
    line-dobj = 'TEXT1'.
    append line to itab.
    line-para = 'P2'.
    line-dobj = 'TEXT2'.
    append line to itab.
    free memory id 'TD'.
    export itab to memory id 'TD'.
    Program2:
    types:
      begin of tab_type,
        para type string,
        dobj type string,
      end of tab_type.
    data:
      id    type c length 10 value 'TEXTS',
      text1 type string value `IKE`,
      text2 type string value `TINA`,
      line  type tab_type,
      itab  type standard table of tab_type,
        itab1  type standard table of tab_type.
    refresh itab.
    import itab from memory id 'TD'.
    free memory id 'TD'.
    clear line.
    loop at itab into line.
      write: / line-para, line-dobj.
      clear line.
    endloop.
    Thanks,
    Jyothi

    Thanks for your links Venkat.
    My problem is solved by using the SHARED BUFFER.
    Here is the code i used inthe first program
            DATA: wa_invdata(10) TYPE c VALUE 'INVDATA'.
            DATA: wa_month(10) TYPE c VALUE 'MONTH'.
            DATA: wa_process(10) TYPE c VALUE 'PROCESS'.
            EXPORT wa_month_end TO MEMORY ID 'MONTHEND'.
            EXPORT e_process = it_inv_data[]
            TO SHARED BUFFER indx(st) ID  wa_process.
            EXPORT e_month = wa_month_end
            TO SHARED BUFFER indx(st) ID  wa_month.
            EXPORT e_invdata = it_tab[]
                TO SHARED BUFFER indx(st) ID wa_invdata.
         EXPORT it_inv_data TO MEMORY ID 'PROCESS'.
            DATA: number           TYPE tbtcjob-jobcount,
                  name             TYPE tbtcjob-jobname VALUE 'CRDR_ ',
                  print_parameters TYPE pri_params.
            CONCATENATE name
                    sy-datum
                    INTO name .
    *CALL FUNCTION 'GET_PRINT_PARAMETERS'
    EXPORTING
       no_dailog              = 'X'
       archive_mode           = '3'
    IMPORTING
       out_parameters         = print_parameters
       out_archive_parameters = archi_parameters
       valid                  = valid_flag
    EXCEPTIONS
       invalid_print_params   = 2
       OTHERS                 = 4.
            CALL FUNCTION 'JOB_OPEN'
              EXPORTING
                jobname          = name
              IMPORTING
                jobcount         = number
              EXCEPTIONS
                cant_create_job  = 1
                invalid_job_data = 2
                jobname_missing  = 3
                OTHERS           = 4.
            IF sy-subrc = 0.
              print_parameters-pdest = 'locl'.
              SUBMIT zsd_crdr_monthend_back_process
                      TO SAP-SPOOL
                      SPOOL PARAMETERS print_parameters
                      WITHOUT SPOOL DYNPRO
                      VIA JOB name NUMBER number
                    AND RETURN.
              IF sy-subrc = 0.
                CALL FUNCTION 'JOB_CLOSE'
                  EXPORTING
                    jobcount             = number
                    jobname              = name
                    strtimmed            = 'X'
                  EXCEPTIONS
                    cant_start_immediate = 1
                    invalid_startdate    = 2
                    jobname_missing      = 3
                    job_close_failed     = 4
                    job_nosteps          = 5
                    job_notex            = 6
                    lock_failed          = 7
                    OTHERS               = 8.
    Second Program
    Program zsd_crdr_monthend_back_process.
      DATA: wa_invdata(10) TYPE c VALUE 'INVDATA'.
      DATA: wa_month(10) TYPE c VALUE 'MONTH'.
      DATA: wa_process(10) TYPE c VALUE 'PROCESS'.
      IMPORT e_invdata = it_tab[]
      FROM SHARED BUFFER indx(st) ID  wa_invdata.
      IF sy-subrc NE 0 OR it_tab[] IS INITIAL..
       No data found for the Month end processing
        MESSAGE s398(00)
           WITH 'No Data Found'.
        STOP.
      ELSE.
    found the data
        IMPORT e_process = it_inv_data[]
        FROM SHARED BUFFER indx(st) ID  wa_process.
        IMPORT e_month = wa_month_end
        FROM SHARED BUFFER indx(st) ID  wa_month.
    Thanks for your help.

  • Portal export/import  on database 8.1.7 on an other machine

    Hello,
    I'm running Oracle8i on AIX. I have problem when I try full export/import my database (version 8.1.7) on an other machine. I have a problem with import users PORTAL30 and PORTAL30_SSO and CREATE TABLE 'WWSEC_ENABLER_CONFIG_INFO$'
    The log when import full is:
    IMP-00017: following statement failed with ORACLE error 22973:
    "CREATE TABLE "WWSEC_ENABLER_CONFIG_INFO$" OF "SEC_ENABLER_CONFIG_TYPE" OID "
    "'BA6473B86EF248BEE0330A18780F48BE' ( PRIMARY KEY ("LS_LOGIN_URL")) OBJECT "
    "ID PRIMARY KEY PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 LOGGING STORAG"
    "E(INITIAL 131072 NEXT 131072 MINEXTENTS 1 MAXEXTENTS 4096 PCTINCREASE 0 FRE"
    "ELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "TBS_PORTAL""
    IMP-00003: ORACLE error 22973 encountered
    ORA-22973: size of object identifier exceeds maximum size allowed
    Thank you.
    Borjana Zekic

    When using export/import between different releases, you need to use the older version of export to create your dump file. In your case, you must use the 8.0.5 export utility to export from your 8.1.7 database. You need a tnsname in your 8.0.5 tnsnames.ora that points to your 8.1.7 database. Then you use SQL Net to connect and do the export. So, instead of doing exp system/manager you would do exp system/manager@tnsname. Just be sure to use the 8.0.5 export, and your import will work fine.

  • Issue with Contract Import

    Hello Folks,
    I am facing an issue with Contract Import. The contract import file has been struck in Contract Document importer and from past 4 days the status is showing as Running. Even in trace file, it is showing as "Mon Mar 16 18:05:51 CET 2015          Error during action"
    Is there any way where the import can be stopped or deleted?
    Any help is highly appreciated!!!
    Thanks in Advance!
    Regards,
    Vignesh

    Thanks John and Peter for your support. Unfortunately I cannot provide a screenshot of the issue since of the confidential material and has to re-do the work. I have re-done all the excel correction and re-created the Indesign merge document on a PC laptop. So far 'no' problems. I have a strong suspicion that the cause has been a macro used in excel to convert text or another one that I used to add values to the colum that InDesign found issue. Definetly InDesign should have reported a problem before merging the data. Yes, I am using CS5.5 with the latest update and Excel 2010 on a Mac. Still, I would like to find out what means the letter 'A' in the Data Merge Panel, since this may have directed me to find why the multiple 'AAA...' shown in the first line in the merge panel. Again thanks John and Peter: much appreciated.

  • Compatibility issue with developer 6i using (64 bit) OS

    Hi All,
    I am using oracle 9iR2 (32 bit) on Linux Red Hat (32 bit) with developer 6i. Now i want to upgrade OS Red Hat (64 bit) and migrate database 10gR2 (64 bit).Is there any compatibility issue with developer 6i using 64 bit Red Hat OS / Oracle Database ?
    Thanks

    Is there any compatibility issue with developer 6i using 64 bit Red Hat OS / Oracle Database ?
    No, as long as developer 6i is at least patch 16. We use it.

  • Export/Import of Database to increase the number of datafiles

    My BW system is close to 400 GB and it only has 3 datafiles.  Since we have 8 CPU cores, I need 8 datafiles.  What is the best way to export/import the database in order to achive 8 datafiles?

    With a BW system that size you can probably get away with it.  You most likely do at least a few full loads so all that data will be evenly distrubuted when you drop and reload.  If you can clean up some of your PSAs and log tables you can probably shrink the 3 files down a little anyway.  If you do little maintenance like that every few weeks, after locking auto growth, you can probably shave 2-3 GBs off each file each week.  Do that for a few months and your large files will love 20 GBs while the other ones start to grow.  Rebuilding indexes also helps with that.  You will be surprised how fast they will level out that way.
    With regard to performance you should be fine.  I certainly wouldnt do it at 10 am on Tuesday though :-).  You can probably get away with it over a weekend though.  It will take basically no time at all to create them and very little IO.  If you choose to clean out the 3 existing ones that will take some time.  I have found it takes about 1-3 hours to shrink a 150 GB datafile down to 100 GBs, that was with 8 CPUs, 30 GBs of RAM, and an SAN that I don't fully understand

  • Issues with export with upgrade to 3.3

    I am experiencing issues with export with 3.3
    The image looks great in the Browser after editing. However when I checked the export folder I discovered corrupt files.
    I only dicovered it because I tried to email one of the images.
    2 out of 200 the blacks had solarized meaning all blacks went to white.
    Never had this problem before in any of my Aperture Versions since day one.
    The bottom line problem how can trust that the images are ok on export whitout having to look at every single export later.
    This is embarissing since I had already delivered a project not knowing there were bad files.
    Additionally images do not seem to be as sharp before the upgrade. I am using Nikon 300 and Nikon 7000 with San Disk Cards.
    Do I have to do some camera profiling ?

    Have you checked the Color Profile setting in the export preset you are using?
    I found my Color Profile settings switched to different settings after the upgrade.
    Try the Generic RGB profile.
    Regards
    Léonie

  • Issue with export in Excel in WAD

    Hi
    I have an issue with export to excel in WAD. I have a web template with two or more tabs.
    On the webtemplate I have two buttons to export to excel. One to export the content from the active tab and one button to export content in all tabs.
    My problem is that when I export to excel it always add the current/active content twice. I've tried with both buttons and they work as expected except that they export one table to much (the active).
    can anyone help / give a clue what to look for.
    br,
    John
    Edited by: John Dolph on Apr 15, 2010 10:50 AM

    Hi
    Sorry I has been a while but the problem still occurs,
    I've defined my export button group like this in a seperate template and I expected that it would only return the resultset from TAB_A and TAB_B but it also export the resultset from the tab I'm currently on.
    <bi:TEMPLATE_PARAMETERS name="TEMPLATE_PARAMETERS" />
                <bi:BUTTON_GROUP_ITEM name="TAB_BUTTONs" designheight="23" designwidth="300" >
                    <bi:BUTTON_LIST type="ORDEREDLIST" >
                        <bi:BUTTON type="COMPOSITE" index="2" >
                            <bi:CAPTION value="EXCEL" />
                            <bi:TOOLTIP value="Table, Graph and/or Table/Graph download to PDF" />
                            <bi:ACTION type="CHOICE" value="INSTRUCTION" >
                                <bi:INSTRUCTION >
                                    <bi:EXPORT >
                                        <bi:ITEM_REF_LIST type="UNORDEREDLIST" >
                                            <bi:ITEM_REF index="1" value="TAB_A_ANALYSIS" />
                                            <bi:ITEM_REF index="2" value="TAB_A_CHART" />
                                            <bi:ITEM_REF index="3" value="TAB_B_ANALYSIS" />
                                            <bi:ITEM_REF index="4" value="TAB_B_CHART" />
                                        </bi:ITEM_REF_LIST>
                                        <bi:EXPORT_FORMAT value="XLS" />
                                    </bi:EXPORT>
                                </bi:INSTRUCTION>
                            </bi:ACTION>
                        </bi:BUTTON>

  • Export INTERNAL TABLE to shared buffer

    Hi all,
    My requirement:
    Export INTERNAL TABLE to shared buffer or SAP Memory.
    Any help will be appreciated.
    Can SET/GET parameter be adopted for internal tables?
    Thanks,
    Tabraiz

    EXPORT (OBJ_TAB) TO MEMORY ID 'ABCD'
    also refer to
    http://help.sap.com/saphelp_45b/helpdata/en/34/8e73a36df74873e10000009b38f9b8/content.htm

  • A project which specifies SQL Server 2012 as the target platform may experience compatibility issues with Microsoft Azure SQL Database.

    Hello
    When I try to publis my asp.net application in azure, i have the next error:
    >C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v12.0\Web\Microsoft.Web.Publishing.targets(4253,5): Warning : A project which specifies SQL Server 2012 as the target platform may experience compatibility issues with Microsoft Azure SQL Database.
    >C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v12.0\Web\Microsoft.Web.Publishing.targets(4253,5): Warning : The project and target databases have different collation settings. Deployment errors might occur.

    Hello
    When I try to publis my asp.net application in azure, i have the next error:
    >C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v12.0\Web\Microsoft.Web.Publishing.targets(4253,5): Warning : A project which specifies SQL Server 2012 as the target platform may experience compatibility issues with Microsoft Azure SQL Database.
    >C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v12.0\Web\Microsoft.Web.Publishing.targets(4253,5): Warning : The project and target databases have different collation settings. Deployment errors might occur.
    Hello,
    This forum is to discuss vb.net issues, your issue is mainly regarding Azure SQL Database, I found that you have posted another thread in that forum
    https://social.msdn.microsoft.com/Forums/azure/en-US/f435921a-14d1-4441-8b2b-32ba3d937aea/a-project-which-specifies-sql-server-2012-as-the-target-platform-may-experience-compatibility-issues?forum=ssdsgetstarted#f435921a-14d1-4441-8b2b-32ba3d937aea.
    I would recommend you focus on that thread to get help.
    Regards,
    Carl
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • System copy with Export/Import Distribyted system UNIX database, Windows CI

    HI
    we have a distributed system witch I want to make a copy of
    Oracle Database on HP-UX
    CI + SCS on Windows
    Dual stack system  ABAP/Java
    1)  I have made an system copy with data base copy tools, and that works fine.
    2)  I have made a system copy export/import of a "central system"
    If I try to make an export/import system copy on a distributed system it wont work.
    ( this work fine in a central system) I dont get anny error messages all gren checks marks ! 
    It seems like the problem are in not finding the profiles from CI ?
    It find the profiles for the sapgateway.
    I someone are more interested I have more documentation with screen shots in a word document.
    I Have used Installation Master SR3 and tryed with some older versions as well.
    I have read SAP Note 970518 - Hom./Het.System Copy SAP NetWeaver 7.0 (2004s) SR2
    SAPINST   
    Using virtual hosts for central and/or DB instances
    SYMPTOM:
    You are using virtual host names for your central instance and/or the database instance (e.g. in a HA environment). Although SAPinst finishes successfully, some export steps are not being performed.
    SOLUTION:
    Start the export with the property SAPINST_USE_HOSTNAME=<virtual host>.
    If the central instance and the database instance run on different hosts, run the database export with SAPINST_USE_HOSTNAME=<virtual host name of the DB server> and - if you run a Java or ABAP+Java system - run the central instance export with SAPINST_USE_HOSTNAME=<virtual host nameof the central instance host>.
    If you run a central system with different host names in the profiles, handle the export as a distributed export, calling SAPinst with parameter SAPINST_USE_HOSTNAME. Perform the export of the database instance on the DB server, calling SAPinst with SAPINST_USE_HOSTNAME=<virtual host of the DB server>.

    Hi I have received folloving answer from SAP:
    You're doing everything correct - besides one thing.
    When you specify the profiles during the database export it cannot be
    correct to specify the profiles of the gateway instance located on
    the same host.
    As a consequence - it picks up the wrong instance G01 and sets the
    property 'hasABAP' and 'hasJava' to 'false'...  - thus the export
    is empty. You can see this within the corresponding 'sapinst.log' file.
    The correct procedure would be to specify the profiles of the central
    instance on the Windows Machine. If you cannot access them, you should
    make them available by Samba or by any type of software which makes
    it's possible to export filesystems or directories from windows to
    unix.
    Best regards
    Bent

Maybe you are looking for