EDT Cat. 15 Data loaded to SAP but failed when updating

Hi,
I have an structure to load data into BP using KCLJ with Cat 15.
Here it is.
AKTYP
TYPE
PARTNER
ROLE1
KUNNR
KUNNR_EXT
BU_GROUP
FIBUKRS
CHIND_ADDR
NAME_CO
NAME_ORG1
NAME_ORG2
STREET
STR_SUPPL1
STR_SUPPL2
STR_SUPPL3
LOCATION
HOUSE_NUM1
POST_CODE1
CITY1
COUNTRY
REGION
LANGU
CHIND_TEL
TEL_NUMBER
TEL_EXTENS
We had the data loaded to SAP with ROLE 000000 and TR0100 but somehow when we tried to update the address thru ROLE TR0100, we had a dump and failed to update. However, the update is good when we use ROLE 000000 to update.
Anyone would have a clue of this?
Thanks

I am new to EDT also. I know Cat 15 which can load BP. SAP has more categories which can load more using KCLJ. Here's my experience.
Under Tcode SIMGH, find IMG structure External Data Transfer for SAP Banking. Under that structure, you can find Display Required and Optional Entry Fields for SEM Banking. When you enter 15 in the Category box, you'll see a whole list of fields about BP. You can use these fields to create your own structure.
After you created your structure, use Define Sender Structure under External Data Transfer for SAP Banking to define your structure. After that, it is done. You can try use KCLJ to load your BP.
If you still have other issues, mostly will be the configuration.
Enjoy.

Similar Messages

  • Data load into SAP ECC from Non SAP system

    Hi Experts,
    I am very new to BODS and I have want to load historical data from non SAP source system  into SAP R/3 tables like VBAK,VBAP using BODS, Can you please provide steps/documents or guidelines on how to achieve this.
    Regards,
    Monil

    Hi
    In order to load into SAP you have the following options
    1. Use IDocs. There are several standard IDocs in ECC for specific objects (MATMAS for materials, DEBMAS for customers, etc., ) You can generate and send IDocs as messages to the SAP Target using BODS.
    2. Use LSMW programs to load into SAP Target. These programs will require input files generated in specific layouts generated using BODS.
    3. Direct Input - The direct input method is to write ABAP programs targetting on specific tables. This approach is very complex and hence a lot of thought process needs to be applied.
    The OSS Notes supplied in previous messages are all excellent guidance to steer you in the right direction on the choice of load, etc.,
    However, the data load into SAP needs to be object specific. So targetting merely the sales tables will not help as the sales document data held in VBAK and VBAP tables you mentioned are related to Articles. These tables will hold sales document data for already created articles. So if you want to specifically target these tables, then you may need to prepare an LSMW program for the purpose.
    To answer your question on whether it is possible to load objects like Materials, customers, vendors etc using BODS, it is yes you can.
    Below is a standard list of IDocs that you can use for this purpose to load into SAP ECC system from a non SAP system.
    Customer Master - DEBMAS
    Article Master - ARTMAS
    Material Master - MATMAS
    Vendor Master - CREMAS
    Purchase Info Records (PIR) - INFREC
    The list is endless.........
    In order to achieve this, you will need to get the functional design consultants to provide ETL mapping for the legacy data to IDoc target schema and fields (better to ahve sa tech table names and fields too). You should then prepare the data after putting it through the standard check table validations for each object along with any business specific conversion rules and validations applied. Having prepared this data, you can either generate flat file output for load into SAP using LSMW programs or generate IDoc messages to the target SAPsystem.
    If you are going to post IDocs directly into SAP target using BODS, you will need to create a partner profile for BODS to send IDocs and define the IDocs you need as inbound IDocs. There are few more setings like RFC connectivity, authorizations etc, in order for BODS to successfully send IDocs into the SAP Target.
    Do let me know if you need more info on any specific queries or issues you may encounter.
    kind regards
    Raghu

  • Unit test runs perfectly fine with NUnit but fails when run from TestExplorer

    Hello all,
    I have a TestProject, Harmony.Tests. In there, I have a method AddApplicationEvent()
    which calls another method Send(InvokeRequestMessage requestMessage) which calls a webservice (OperationHandlerBrokerWebService). 
    The code snippet looks like this. This is not the complete code but a part where we are calling the web service. It fails on the underlined Italic line of code.
    OperationHandlerBrokerWebService brokerService = new OperationHandlerBrokerWebService();
    brokerService.UseDefaultCredentials = true;
    brokerService.Url = address;
    brokerService.Timeout = timeoutInMilliseconds;
    byte[] serializedResponseMessage = brokerService.InvokeOperationHandler(serializedRequestMessage);
    The same test works and passed fine when I ran it with NUnit, but failed with following exception when I tried to run it from TestExplorer. 
    Test Name: AddApplicationEvent
    Test FullName: N4S.Harmony.Tests.CaseManagement.HarmonyFacadeTests.AddApplicationEvent
    Test Source: d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\CaseManagement\HarmonyFacadeTests.cs : line 665
    Test Outcome: Failed
    Test Duration: 0:00:00.296
    Result Message:
    SetUp : Message returned System.Web.Services.Protocols.SoapException: Server was unable to process request. ---> System.Reflection.TargetInvocationException: Exception has been thrown by the target of an invocation. ---> System.ArgumentException: Invalid token for impersonation - it cannot be duplicated.
    at System.Security.Principal.WindowsIdentity.CreateFromToken(IntPtr userToken)
    at System.Security.Principal.WindowsIdentity..ctor(SerializationInfo info)
    at System.Security.Principal.WindowsIdentity..ctor(SerializationInfo info, StreamingContext context)
    --- End of inner exception stack trace ---
    at System.RuntimeMethodHandle._SerializationInvoke(Object target, SignatureStruct& declaringTypeSig, SerializationInfo info, StreamingContext context)
    at System.Reflection.RuntimeConstructorInfo.SerializationInvoke(Object target, SerializationInfo info, StreamingContext context)
    at System.Runtime.Serialization.ObjectManager.CompleteISerializableObject(Object obj, SerializationInfo info, StreamingContext context)
    at System.Runtime.Serialization.ObjectManager.FixupSpecialObject(ObjectHolder holder)
    at System.Runtime.Serialization.ObjectManager.DoFixups()
    at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(HeaderHandler handler, __BinaryParser serParser, Boolean fCheck, Boolean isCrossAppDomain, IMethodCallMessage methodCallMessage)
    at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, HeaderHandler handler, Boolean fCheck, Boolean isCrossAppDomain, IMethodCallMessage methodCallMessage)
    at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream)
    at N4S.Forms.OperationHandlerBroker.AMessage.DeserializeMessage(Byte[] serializedMessage)
    at N4S.Forms.OperationHandlerBroker.WebServiceServer.BrokerService.InvokeOperationHandler(Byte[] serializedInvokeRequestMessage)
    --- End of inner exception stack trace ---
    expected: <0>
    but was: <1>
    Result StackTrace:
    at N4S.Harmony.Tests.TestHelper.InvokeOperation(OperationHandler handler, OperationHandlerInput input, Boolean expectedToWork) in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\TestHelper.cs:line 136
    at N4S.Harmony.Tests.TestHelper.LoginAsUser(String username, String password) in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\TestHelper.cs:line 394
    at N4S.Harmony.Tests.TestHelper.Login(TestUserName requestedUser) in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\TestHelper.cs:line 377
    at N4S.Harmony.Tests.TestHelper.LoginAsAdvisor() in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\TestHelper.cs:line 230
    at N4S.Harmony.Tests.CaseManagement.HarmonyFacadeTests.Login() in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\CaseManagement\HarmonyFacadeTests.cs:line 76
    at N4S.Harmony.Tests.CaseManagement.HarmonyFacadeTests.SetupTest() in d:\TFS\TMW\Dev\TMWOnline\Harmony\N4S.Harmony.Tests\CaseManagement\HarmonyFacadeTests.cs:line 67
    I am not sure what causing the issue. I checked the Credentials, Windows Identity during both the test run and there is no difference. Please advise !!
    Thanks,
    Deepak

    Hi Tina,
    Thanks for your reply. 
    I do have NUnit adapter installed. I even noticed that the test runs fine with NUnit GUI and also if I run it through Resharper Test Explorer window. 
    As you can see in the image above the same test is passed when I ran it from Resharper Unit Test Explorer window but fails when I ran it from Test Explorer window. I also captured the information on fiddler. 
    There was a significant difference in the Header Content length. Also under the User-Agent property the protocol versions are different. 
    Not sure why VSTest ExecutionEngine is picking a different version.
    The UnitTest in question is calling a webservice method which in turn calls a method from another referenced project. 
    Web Service class
    using System;
    using System.Web.Services;
    using N4S.Forms.OperationHandlerBroker.Server;
    using NLog;
    namespace N4S.Forms.OperationHandlerBroker.WebServiceServer
    /// <summary>
    /// The operaton-handler broker service.
    /// </summary>
    [WebService(Description = "The N4S Forms Operation-Handler Broker Web-Service.", Name = "OperationHandlerBrokerWebService",
    Namespace = "N4S.Forms.OperationHandlerBroker.WebServiceServer")]
    [WebServiceBinding(ConformsTo = WsiProfiles.BasicProfile1_1)]
    public class BrokerService : WebService
    { /// <summary>
    /// Calls <see cref="HandleRequest"/>. Updates performance-counters.
    /// </summary>
    /// <param name="serializedInvokeRequestMessage">the binary-serialized <see cref="InvokeRequestMessage"/></param>
    /// <returns>the binary-serialized response message</returns>
    [WebMethod(BufferResponse = true, CacheDuration = 0, Description = "Invokes the requested operation-handler and returns a binary-serialized response-message.", EnableSession = false)]
    public byte[] InvokeOperationHandler(byte[] serializedInvokeRequestMessage)
    logger.Trace(Strings.TraceMethodEntered);
    PerformanceMonitor.RecordRequestStarted();
    InvokeRequestMessage requestMessage = (InvokeRequestMessage) AMessage.DeserializeMessage(serializedInvokeRequestMessage);
    InvokeResponseMessage responseMessage;
    try
    responseMessage = HandleRequest(requestMessage);
    PerformanceMonitor.RecordSuccessfulRequest();
    catch (Exception)
    PerformanceMonitor.RecordFailedRequest();
    throw;
    finally
    PerformanceMonitor.RecordRequestEnded();
    logger.Trace(Strings.TraceMethodExiting);
    return AMessage.SerializeMessage(responseMessage);
    UnitTest snippet
    OperationHandlerBrokerWebService brokerService = new OperationHandlerBrokerWebService();
    brokerService.UseDefaultCredentials = true;
    byte[] serializedResponseMessage = brokerService.InvokeOperationHandler(serializedRequestMessage);
    Please advise.
    Thanks,
    Deepak

  • SSIS package works in development environment but fails when job scheduler executes, file path invalid

    SIS package works in development environment but fails when job scheduler executes, file path invalid
    Relatively simple package to get OLE-DB connection to MS FoxPro 9.0 DB
    The failure in the job log states that the path is invalid its a network path (\\192.168.1.xxx\foldername) this has been run several ways with the remote computer being mapped as a network drive and through the \\ notation described above.
    Thinking it was a security issue between the SQL agent account and my account I tested by subsitution myself as a proxy account for the agent when running this job, again same result failed on network path.
    One issue I see is that the remote computer is running Server 2000 (legacy software incompatable with newer versions) 
    Is it possible that this is a security issue, since if i understand correctly the current MS domain security model didn't exist until Server 2003.

    Hi REIData,
    Have you got the issue resolved? Based on your description, please make sure the target folder is shared properly. If the computer on which the SQL Server Agent job runs doesn’t join the domain as the server that hosts the shared folder, you have to share
    the folder with everyone by adding “Everyone” to the people list on the File Sharing page of the folder and assign "Read/Write" permission to it.
    Regards,
    Mike Yin
    TechNet Community Support

  • NTRealm failed when updating to SP8

    I tried to update NTRealm from SP4 to SP8. I followed all the instructions from
    document. When I restart weblogic the following error message comes up:
    Unable to initialize server: java.lang.RuntimeException: OS error=53: T
    fatal initialization exception
    java.lang.RuntimeException: OS error=53: T
    at weblogic.security.ntrealm.NTResumeEnum.populate0(Native Method)
    at weblogic.security.ntrealm.NTResumeEnum.populate(NTDelegate.java, Comp
    iled Code)
    at weblogic.security.ntrealm.NTResumeEnum.<init>(NTDelegate.java, Compil
    ed Code)
    at weblogic.security.ntrealm.NTDelegate.getGroupNames(NTDelegate.java:16
    7)
    at weblogic.security.ntrealm.NTRealm.getGroup(NTRealm.java, Compiled Cod
    e)
    at weblogic.security.acl.CachingRealm.getGroup(CachingRealm.java:1435)
    at weblogic.security.acl.CachingRealm.getPrincipal(CachingRealm.java:102
    0)
    at weblogic.security.acl.CachingRealm.addPermission(CachingRealm.java,
    C
    ompiled Code)
    at weblogic.security.acl.CachingRealm.setupAcls(CachingRealm.java, Compi
    led Code)
    at weblogic.security.acl.CachingRealm.<init>(CachingRealm.java:706)
    at weblogic.security.acl.CachingRealm.<init>(CachingRealm.java:564)
    at weblogic.t3.srvr.T3Srvr.initializeSecurity(T3Srvr.java:1759)
    at weblogic.t3.srvr.T3Srvr.start(T3Srvr.java, Compiled Code)
    at weblogic.t3.srvr.T3Srvr.main(T3Srvr.java:827)
    at java.lang.reflect.Method.invoke(Native Method)
    at weblogic.Server.startServerDynamically(Server.java:99)
    at weblogic.Server.main(Server.java:65)
    at weblogic.Server.main(Server.java:55)
    Thank you very much.

    Sorry, this is a follow-up to the following posting from April:
    Subject: NTRealm failed when updating to SP8
    Date: Thu Apr 19 11:54:03 PDT 2001
    From: "Bill" <[email protected]>
    Organization: BEA SYSTEMS Inc
    Newsgroups: [weblogic.developer.interest.security]
    I tried to update NTRealm from SP4 to SP8. I followed all the instructions
    from
    document. When I restart weblogic the following error message comes up:
    Unable to initialize server: java.lang.RuntimeException: OS error=53: T
    fatal initialization exception
    java.lang.RuntimeException: OS error=53: T
    at weblogic.security.ntrealm.NTResumeEnum.populate0(Native Method)
    at weblogic.security.ntrealm.NTResumeEnum.populate(NTDelegate.java,
    Comp
    iled Code)
    at weblogic.security.ntrealm.NTResumeEnum.<init>(NTDelegate.java,
    Compil
    ed Code)
    at
    weblogic.security.ntrealm.NTDelegate.getGroupNames(NTDelegate.java:16
    7)
    at weblogic.security.ntrealm.NTRealm.getGroup(NTRealm.java, Compiled
    Cod
    e)
    at
    weblogic.security.acl.CachingRealm.getGroup(CachingRealm.java:1435)
    at
    weblogic.security.acl.CachingRealm.getPrincipal(CachingRealm.java:102
    0)
    at
    weblogic.security.acl.CachingRealm.addPermission(CachingRealm.java,
    C
    ompiled Code)
    at weblogic.security.acl.CachingRealm.setupAcls(CachingRealm.java,
    Compi
    led Code)
    at weblogic.security.acl.CachingRealm.<init>(CachingRealm.java:706)
    at weblogic.security.acl.CachingRealm.<init>(CachingRealm.java:564)
    at weblogic.t3.srvr.T3Srvr.initializeSecurity(T3Srvr.java:1759)
    at weblogic.t3.srvr.T3Srvr.start(T3Srvr.java, Compiled Code)
    at weblogic.t3.srvr.T3Srvr.main(T3Srvr.java:827)
    at java.lang.reflect.Method.invoke(Native Method)
    at weblogic.Server.startServerDynamically(Server.java:99)
    at weblogic.Server.main(Server.java:65)
    at weblogic.Server.main(Server.java:55)
    Thank you very much.

  • Mass data load into SAP R/3 - with XI?

    Hi guys!
    I have an issue - mass data migration into SAP R/3. Is XI a good solution? It will be about 60GB of data. Or is there a better way of this data load?
    Thanx a lot!
    Olian

    hi,
    SAP doesn't recomment using XI for mass data migration
    and 60 Gb is certainly too much
    use LSMW for that purpose
    Regards,
    michal

  • Data Load process for 0FI_AR_4  failed

    Hi!
    I am aobut to implement SAP Best practices scenario "Accounts Receivable Analysis".
    When I schedule data load process in Dialog immediately for Transaction Data 0FI_AR_4 and check them in Monitor the the status is yellow:
    On the top I can see the following information:
    12:33:35  (194 from 0 records)
    Request still running
    Diagnosis
    No errors found. The current process has probably not finished yet.
    System Response
    The ALE inbox of BI is identical to the ALE outbox of the source system
    or
    the maximum wait time for this request has not yet been exceeded
    or
    the background job has not yet finished in the source system.
    Current status
    No Idocs arrived from the source system.
    Question:
    which acitons can  I do to run the loading process succesfully?

    Hi,
    The job is still in progress it seems.
    You could monitor the job that was created in R/3 (by copying the technical name in the monitor, appending "BI" to is as prefix, and searching for this in SM37 in R/3).
    Keep on eye on ST22 as well if this job is taking too long, as you may have gotten a short dump for it already, and this may not have been reported to the monitor yet.
    Regards,
    De Villiers

  • Data Load with warnings, but dataload.err empty

    Hi guys,
    laoding data on EAS Console for a Essbase database gives below message:
    Database has been cleared before loading.
    dataload.err is EMPTY.
    Any ideas??
    Some data is loaded but how can I find out where it stopped or is all my data loaded?
    Message:
    Parallel dataload enabled: [1] block prepare threads, [1] block write threads.
    Transaction [ 0x190004( 0x4fd89600.0x64578 ) ] aborted due to status [1003050].
    There were errors, look in H:\My Documents\temp\dataload.err
    Database import completed ['AXN_Sim'.'BewMod']
    Output columns prepared: [0]
    Thanks Bernd

    Hi,
    There is nothing much you can say whether the data is completely loaded or not. You have to check by yourself on that.
    Since the loading was aborted with a warning, it could be a network issue..
    Try to load a small file and then check whether you are facing the problem or you coulfd also post your finding on that..
    Thanks,
    CM

  • Data loaded into ODS, but no activation possible

    Hello Experts,
    I've data loaded data into an ODS, but no activation is possible!
    In Monitor it just say something like "not activated yet".
    usually it only takes a minute to activate, so I'm confised.
    any suggestions?
    Thanx
    Axel

    Hi,
    In SM37 -> give Job Name as BI_ODSA* and USer as  *  and check is there any job with Active status. If so, We can say ODS Activation is on process.
    With rgds,
    Anil Kumar Sharma .P

  • Data load erroring out but giving a completed status

    Hello Everyone,
    We are on BPC 5.1 SP3. We have our data imported daily on a schedule. The actuals import without any errors or problems. But our Budget/Forecast file is not importing correctly. We have it set to run each day at 5:30 am. The files are placed on the server and the scheduled job grabs the files imports them with the conversion and transformation files. It has been running successfully for months. Now all of a sudden they stopped working correctly. When you select View Status to see what happened. The package says it's completed but if you click on details. It shows that the data conversion on step 2 failed. Since it doesn't say failed we don't get a reject list or anything that we can use to troubleshoot why it's failing. Both the actual and forecast files use the same transformation and conversion files.  I have revalidated the transformation and conversion files with both the data files and didn't get any errors or kickouts with the data files. So I am at a loss as to what would be causing the failure for one data file but not the other.
    I forgot to add that when we run the import package manually with the forecast data file it runs successfully.
    Any help would be greatly appreciated.
    Thanks,
    Eleasha
    Edited by: Eleasha Counts on Aug 4, 2009 10:33 PM

    Hi,
    Ok I thought you are not using standard package.
    In this case you really have to check your conversion and transformation files.
    Also I suggest to try to use first a file with just online which normally should work and try to import that one.
    If that's working then coversion and transformation file are fine you have to check the format of your input file.
    Probably you have a TAB char instead of blank char or things like that.
    Regards
    Sorin Radulescu

  • Full and delta Data loads in SAP BI 7.0

    Hi,
    Kindly explain me the difference between full and delta data loads in cube and DSO.
    When do we select change log or active table(with or without archive) with full/delta in DTP for DSO.
    Please explain the differences in data load for above different combinations.
    With thanks and regards,
    Mukul

    Please search the forums. This topic has already been discussed a lot of times.

  • Data loaded at YTD, but showing Periodic in HFM until re-Export, then YTD

    We are using FDM 11.1.2.1 (FM11X-G5-E Target adapter) to load YTD data to HFM 11.1.2.1. The "View" dimension in FDM is set to YTD and we don't have any other custom event or calculation scripting happening. Also confirmed the source file values are YTD, as well as the value in Export grid in FDM and the .dat files in the Outbox.
    When we then view the data in an HFM grid or SmartView, many times it displays as Periodic even though the View is set to YTD. We've double checked many HFM settings (UsePVAForFlowAccounts=Y, ZeroViewForNonadj = YTD, ZeroViewForAdj = YTD, ConsolidateYTD = Y, DefFreqForICTrans = Y), and ensured we are consolidating with View = YTD. I can even drill through to FDM on the Periodic value and see the correct YTD value for that POV in FDM.
    Now the strange part: If we then simply re-run Export in FDM, and refresh the grid/SmartView for the same POV, the balance is now showing the correct YTD value. It's seems to show up randomly for different data sources, at different times, and this fix always works. It's almost as if the Periodic value or setting is "stuck", and then gets fixed after another Export from FDM.
    However, the client's long term future process cannot be "re-Export each monthly load twice". Not sure if anyone has encountered this and if it's a result of FDM, HFM, backend DB, patching, metadata, etc. Any help or thoughts would be appreciated. Thanks.

    Yes, I know [YEAR]. I meant, I have never noticed data entry allowed on [YEAR].
    Is the accounts properly intersected with custom dimensions? i.e the Account property of TopCustom1Member,TopCustom2Member...etc. If that is not set then, accounts will not have valid intersection with Customs and you can not enter data.
    Secondly., What is the default frequency for the scenario you are selecting? MTD? Are you looking at a month in Period (Jan/Feb etc) or at Q1/Q2 level?

  • SAP_CMINIT3 : rc=20 Connect to SAP gateway failed when connecting SAP CR

    Dear SAP Gurus,
    I have the follwowing issue when I try to access SAP Crystal Reports 2011 on my desktop.
    I can access the SAP ERP 6.0 back-end system without problem, but when I try to connect to SAP Crystal Reports, as soon as I enter my login details, I get the following error message:
    Logon Failed.
    Details: SAP_CMINIT3 : rc=20 > Connect to SAP gateway failed
    Connect_PM GWHOST=192.12.353.12, GWSERV=sapgw00, SYSNR=00
    LOCATION CPIC (TCP/IP) on local host with Unicode
    ERROR partner 192.12.353.12:3300' not reached
    TIME Thu Oct 27 12:02:17 201
    RELEASE 700
    COMPONENT NI (network interface)
    VERSION 38
    RC -10
    MODULE nixxi.cpp
    LINE 2823
    DETAIL NiPConnect2
    SYSTEM CALL connect
    ERRNO 10060
    ERRNO TEXT WSAETIMEDOUT: Connection timed out
    COUNTER 3
    I can login to SAP Crystal Reports using other desktops, but cannot access using my own. This is the same for other people, they also cannot access using my desktop.
    Thanks in advance,  Marc

    HI Mark,
    Are you and Ramish working together?
    See this post for your answer: Connection keeps failing
    Don

  • Scp works when logging with my user id but fails when using another one

    Hello all:
    I wrote a script that uses scp to transfer specific local files to a server. The script takes the user id and password and pass them as scp's parameters for authentication.
    Everything works fine when I uses my own user id to test the script but it fails when I used another id (with ligitimate password) to do the same thing.
    I suspect it's a privilege issue. Both ids belong to the same group. Where should I look to fingure this out?
    Thanks,
    John

    SCP is part of SSH. SSH requires the exchange of authorization keys to prevent man-in-the-middle attacks.
    When you log in the first time you may see a prompt like:
    RSA key fingerprint is 98:2e:d7:e0:de:9f:ac:67:28:d2:42:2e:37:16:58:45.
    Are you sure you want to continue connecting (yes/no)?
    This might be where your script fails. Since it prompts only the first time, you may just have to do one interactive SSH login to create an entry in ~/.ssh/known_hosts
    Or setup User equivalence to omit the exchange of passwords.
    You may find the following interesting:
    Oracle Linux 5 Howto: Installing and using cURL with support for SCP and SFTP
    OL5 Howto: Installing and using cURL with support for SCP and SFTP
    see 3) Set up SSH user equivalence for SCP and SFTP

  • MPR testing is passed, but failed when upload Result Package to verify.

    Hi,
    I am attempting to certify our application using MPR-Test-Tool v4.6 for Windows Server 2012 R2.
    The MPR tool shows the testing Pass, and generated a .zip Result Package.
    But it's failed when uploads the .zip Result Package to verify in MPR web-site "Test My Apps" page, both Silver and Gold level are failed.
    In the "Test My Apps" page, I couldn't find the reason, error message or some hint for the failure.
    I am appreciate any help,
    Regards.
    Regards,

    Hello,
    Please check if you import the certificate into the Exchange server "Trusted Root Certificates".
    If there is a CA server in your domain, you also need to import your certificate into the trusted root certificates on the CA server.
    Cara Chen
    TechNet Community Support

Maybe you are looking for