Creating data and .par files using SWIFT Integration Package

Hi,
I have a requirement to generate a data and .par file using SAP PI Integration package for SWIFT.
I am following the SAP Note: 1303428.
I have created 1 sender and 2 receiver comm channels(one for payload and other for par). Used operation mapping SWIFT_payload_parFile_split in the Interface determination.
I am using the adapter module : localejbs/swift/FileActConversionToSWIFTModule and setting the parameter DetachParameters to true. This adapter module is being used in all the three channels(1 sender and 2 receivers).
I have used ASMA ans has set the FileName checkbox.
Now after placing the file in the input directory, the file with the same name gets created in the output directory but the file is exactly same and also no .par file is getting created. I have set the Empty file handling to Ignore, so it shows that there is no data to  create a .par file and only payload file is getting created but the payload file is exactly same.
Also if I use the adapter module : localejbs/swift/FileActConversionToSWIFTModule in only the sender communication channel, a payload file gets created like below.
<?xml version="1.0" encoding="UTF-8"?>
-<ns1:SWIFT_payload xmlns:ns1="http://sap.com/xi/SWIFT"><payload>PD94bWwgdmVyc2lvbj0iMS4wIiBlbmNvZGluZz0idXRmLTgiPz4KPEZpbGU+CiA8UGFyYW1ldGVycyB4bWxucz0idXJuOnN3aWZ0OnNhZzp4c2Q6ZnRhLnBhcmFtLjEuMCIgeG1sbnM6eHNpPSJodHRwOi8vd3d3LnczLm9yZy8yMDAxL1hNTFNjaGVtYS1pbnN0YW5jZSI+CiAgPE92ZXJyaWRlcz4KICAgPFJlc3BvbmRlcj5jbj1jZngsb3U9bmEsbz1jaXRpZ2IybCxvPXN3aWZ0PC9SZXNwb25kZXI+CiAgIDxTZXJ2aWNlPnN3aWZ0LmNvcnAuZmEhcDwvU2VydmljZT4KICAgPFJlcXVlc3RUeXBlPnBhaW4uMDAxLjAwMS4wMzwvUmVxdWVzdFR5cGU+CiAgIDxUcmFuc2ZlckRlc2NyaXB0aW9uPkIwNDExMC1CYXRjaDE3ODc8L1RyYW5zZmVyRGVzY3JpcHRpb24+CiAgIDxUcmFuc2ZlckluZm8+QjA0MTEwLUJhdGNoMTc4NzwvVHJhbnNmZXJJbmZvPgogICA8RmlsZURlc2NyaXB0aW9uPjIwNzY4PC9GaWxlRGVzY3JpcHRpb24+CiAgIDxGaWxlSW5mbz5Td0NvbXByZXNzaW9uPW5vbmU8L0ZpbGVJbmZvPgogICA8Tm9uUmVwdWRpYXRpb24+VFJVRTwvTm9uUmVwdWRpYXRpb24+CiAgIDxTaWduPlRSVUU8L1NpZ24+CiAgIDxQcmlvcml0eT5Ob3JtYWw8L1ByaW9yaXR5PgogIDwvT3ZlcnJpZGVzPgogPC9QYXJhbWV0ZXJzPgogPERvY3VtZW50IHhtbG5zPSJ1cm46aXNvOnN0ZDppc286MjAwMjI6dGVjaDp4c2Q6cGFpbi4wMDEuMDAxLjAzIiB4bWxuczp4c2k9Imh0dHA6Ly93d3cudzMub3JnLzIwMDEvWE1MU2NoZW1hLWluc3RhbmNlIj4KICA8Q3N0bXJDZHRUcmZJbml0bj4KICAgPEdycEhkcj4KICAgIDxNc2dJZD4xMDAwMzI4MTE1PC9Nc2dJZD4KICAgIDxDcmVEdFRtPjIwMTQtMDMtMjhUMTk6MjY6Mzc8L0NyZUR0VG0+CiAgICA8TmJPZlR4cz4xPC9OYk9mVHhzPgogICAgPEN0cmxTdW0+NTkwLjAwPC9DdHJsU3VtPgogICAgPEluaXRnUHR5PgogICAgIDxObT5BTEVYSU9OIElOVC4gU0FSTDwvTm0+CiAgICAgPElkPgogICAgICA8T3JnSWQ+CiAgICAgICA8QklDT3JCRUk+QUxYTlVTMjBYWFg8L0JJQ09yQkVJPgogICAgICA8L09yZ0lkPgogICAgIDwvSWQ+CiAgICA8L0luaXRnUHR5PgogICA8L0dycEhkcj4KICAgPFBtdEluZj4KICAgIDxQbXRJbmZJZD4xMDAwMzI4MTE1PC9QbXRJbmZJZD4KICAgIDxQbXRNdGQ+VFJGPC9QbXRNdGQ+CiAgICA8QnRjaEJvb2tnPmZhbHNlPC9CdGNoQm9va2c+CiAgICA8TmJPZlR4cz4xPC9OYk9mVHhzPgogICAgPEN0cmxTdW0+NTkwLjAwPC9DdHJsU3VtPgogICAgPFBtdFRwSW5mPgogICAgIDxJbnN0clBydHk+Tk9STTwvSW5zdHJQcnR5PgogICAgIDxTdmNMdmw+CiAgICAgIDxDZD5TRVBBPC9DZD4KICAgICA8L1N2Y0x2bD4KICAgIDwvUG10VHBJbmY+CiAgICA8UmVxZEV4Y3RuRHQ+MjAxNC0wMy0yOTwvUmVxZEV4Y3RuRHQ+CiAgICA8RGJ0cj4KICAgICA8Tm0+QUxYTiBCRU5FTFVYIEJWIE5MIEJSQU5DSDwvTm0+CiAgICAgPFBzdGxBZHI+CiAgICAgIDxTdHJ0Tm0+U3RyYWF0PC9TdHJ0Tm0+CiAgICAgIDxUd25ObT5OZXRoZXJsYW5kczwvVHduTm0+CiAgICAgIDxDdHJ5Pk5MPC9DdHJ5PgogICAgIDwvUHN0bEFkcj4KICAgICA8Q3RyeU9mUmVzPk5MPC9DdHJ5T2ZSZXM+CiAgICA8L0RidHI+CiAgICA8RGJ0ckFjY3Q+CiAgICAgPElkPgogICAgICA8SUJBTj5OTFhYQU5CQTEyMzAwNDU2NzY3ODkwPC9JQkFOPgogICAgIDwvSWQ+CiAgICAgPENjeT5FVVI8L0NjeT4KICAgIDwvRGJ0ckFjY3Q+CiAgICA8RGJ0ckFndD4KICAgICA8RmluSW5zdG5JZD4KICAgICAgPEJJQz5BQk5BTkwyWFhYWDwvQklDPgogICAgICA8UHN0bEFkcj4KICAgICAgIDxDdHJ5Pk5MPC9DdHJ5PgogICAgICA8L1BzdGxBZHI+CiAgICAgPC9GaW5JbnN0bklkPgogICAgPC9EYnRyQWd0PgogICAgPENocmdCcj5TTEVWPC9DaHJnQnI+CiAgICA8Q2R0VHJmVHhJbmY+CiAgICAgPFBtdElkPgogICAgICA8RW5kVG9FbmRJZD5OTDEyMzAwNDAwMDAwMDwvRW5kVG9FbmRJZD4KICAgICA8L1BtdElkPgogICAgIDxBbXQ+CiAgICAgIDxJbnN0ZEFtdCBDY3k9IkVVUiI+NTkwLjAwPC9JbnN0ZEFtdD4KICAgICA8L0FtdD4KICAgICA8Q2R0ckFndD4KICAgICAgPEZpbkluc3RuSWQ+CiAgICAgICA8QklDPkFCTkFOTFhYWFhYPC9CSUM+CiAgICAgICA8Q2xyU3lzTW1iSWQ+CiAgICAgICAgPE1tYklkPjAwMzwvTW1iSWQ+CiAgICAgICA8L0NsclN5c01tYklkPgogICAgICAgPE5tPkFCTiBBbXJvPC9ObT4KICAgICAgIDxQc3RsQWRyPgogICAgICAgIDxDdHJ5Pk5MPC9DdHJ5PgogICAgICAgPC9Qc3RsQWRyPgogICAgICA8L0Zpbkluc3RuSWQ+CiAgICAgPC9DZHRyQWd0PgogICAgIDxDZHRyPgogICAgICA8Tm0+QUxYTiBOTCBEb21lc3RpYyBWZW5kb3I8L05tPgogICAgICA8UHN0bEFkcj4KICAgICAgIDxDdHJ5Pk5MPC9DdHJ5PgogICAgICA8L1BzdGxBZHI+CiAgICAgIDxJZD4KICAgICAgIDxPcmdJZD4KICAgICAgICA8T3Rocj4KICAgICAgICAgPElkPjAwMTUwMDAxOTc8L0lkPgogICAgICAgIDwvT3Rocj4KICAgICAgIDwvT3JnSWQ+CiAgICAgIDwvSWQ+CiAgICAgPC9DZHRyPgogICAgIDxDZHRyQWNjdD4KICAgICAgPElkPgogICAgICAgPElCQU4+TkwxMjAwMzA0NTY3ODkxMjAwMDA8L0lCQU4+CiAgICAgIDwvSWQ+CiAgICAgIDxDY3k+RVVSPC9DY3k+CiAgICAgIDxObT5BTFhOIE5MIERvbWVzdGljIFZlbmRvcjwvTm0+CiAgICAgPC9DZHRyQWNjdD4KICAgICA8Um10SW5mPgogICAgICA8VXN0cmQ+L1BNREQvVEVTVDY4MSw1OTAuMDAsRVVSLDIwMTQwMzI8L1VzdHJkPgogICAgIDwvUm10SW5mPgogICAgPC9DZHRUcmZUeEluZj4KICAgPC9QbXRJbmY+CiAgPC9Dc3RtckNkdFRyZkluaXRuPgogPC9Eb2N1bWVudD4KPC9GaWxlPgo=</payload></ns1:SWIFT_payload>
But while creating the par file, it shows that the file could not be overwritten and so the .par file is not getting created.
I need to understand that:
1)  How do I configure both of my receiver channels i.e. what should be the difference. Currently, I am just setting the
same Output directory in both and file name schema * and has used ASMA with FileName parameter. So same name files are getting created and so one file gets discarded. How to create a .par file.
2) Also Is the file above is the correct file required by SWIFT.
3) Also when I check in SXMB_MONI, I can see that after running the adapter module same payload as given above is going in both payload and par files i.e. also if I am using different names in my receiver communication channels two files are getting created payload and par both are having the same payload. So what exactly should get created.
Kindly guide on this implementation.

Hi,
I am able to generate the .par file by setting the localsecurity to true and the KeyId from the key manager.
Now two files are getting created .xml and .par. .xml file which is the payload file is the same as the input file with no difference and contains both the overrides and the data parameters. The .par file contains the Algorithm and the Value.
Is this correct?
Also suppose the input file name is SEPA.xml, then the payload file is created with the name as
SEPA.xml and the par file is created with the name SEPA.xml.par. I need only SEPA.par. How to achieve this.

Similar Messages

  • Why can't I change the Created Date on a file using Terminal?

    I bought a used Contour helmet camera to take pictures of the kids while they're skiing, and I cannot access the camera itself to change the clock in it (USB port is broken).
    I get videos from it just fine, but every time I remove the battery the date/time on the camera resets to January 2008, so my skiing videos end up someplace completely different from what I expect in iMovie.
    I thought I could change the Created Date with Touch in Terminal, but it only handles the Modified Date.  I know I can purchase File Date Changer 5 from the App Store to do this, but I'd like to know why I can't do it in Terminal.
    Anybody know?  Or know of a different way that I COULD do it in Terminal?

    pessex wrote: Anybody know?  Or know of a different way that I COULD do it in Terminal?
    Don't know about Terminal, but, because you are using 10.9.x and iMovie, try the iMovie method:

  • How to get the most current file based on date and time stamp using SSIS?

    Hello,
    Let us assume that files get copied in a specific directory. We need to pick up a file and load data. Can you guys let me know how to get the most current file based on date and time stamp using SSIS?
    Thanks
    thx regards dinesh vv

    hi simon
    i excuted this script it is giving error..
       Microsoft SQL Server Integration Services Script Task
       Write scripts using Microsoft Visual C# 2008.
       The ScriptMain is the entry point class of the script.
    using System;
    using System.Data;
    using Microsoft.SqlServer.Dts.Runtime;
    using System.Windows.Forms;
    namespace ST_9a6d985a04b249c2addd766b58fee890.csproj
        [System.AddIn.AddIn("ScriptMain", Version = "1.0", Publisher = "", Description = "")]
        public partial class ScriptMain : Microsoft.SqlServer.Dts.Tasks.ScriptTask.VSTARTScriptObjectModelBase
            #region VSTA generated code
            enum ScriptResults
                Success = Microsoft.SqlServer.Dts.Runtime.DTSExecResult.Success,
                Failure = Microsoft.SqlServer.Dts.Runtime.DTSExecResult.Failure
            #endregion
            The execution engine calls this method when the task executes.
            To access the object model, use the Dts property. Connections, variables, events,
            and logging features are available as members of the Dts property as shown in the following examples.
            To reference a variable, call Dts.Variables["MyCaseSensitiveVariableName"].Value;
            To post a log entry, call Dts.Log("This is my log text", 999, null);
            To fire an event, call Dts.Events.FireInformation(99, "test", "hit the help message", "", 0, true);
            To use the connections collection use something like the following:
            ConnectionManager cm = Dts.Connections.Add("OLEDB");
            cm.ConnectionString = "Data Source=localhost;Initial Catalog=AdventureWorks;Provider=SQLNCLI10;Integrated Security=SSPI;Auto Translate=False;";
            Before returning from this method, set the value of Dts.TaskResult to indicate success or failure.
            To open Help, press F1.
            public void Main()
                string file = Dts.Variables["User::FolderName"].Value.ToString();
                string[] files = System.IO.Directory.GetFiles(Dts.Variables["User::FolderName"].Value.ToString());
                System.IO.FileInfo finf;
                DateTime currentDate = new DateTime();
                string lastFile = string.Empty;
                foreach (string f in files)
                    finf = new System.IO.FileInfo(f);
                    if (finf.CreationTime >= currentDate)
                        currentDate = finf.CreationTime;
                        lastFile = f;
                Dts.Variables["User::LastFile"].Value = lastFile;
                Dts.TaskResult = (int)ScriptResults.Success;
    thx regards dinesh vv

  • How can I create an csv/excel file using pl/sql and then sending that file

    How can I create an csv/excel file using pl/sql and then sending that file to a clients site using pl/sql?
    I know how to create the csv/excel file but I can't figure out how I would get it to the clients site.

    968776 wrote:
    How can I create an csv/excel file using pl/sql and then sending that file to a clients site using pl/sql?
    I know how to create the csv/excel file but I can't figure out how I would get it to the clients site.You are trying to do it at a wrong place..
    Whay do you want database (pl/sql) code to do these things?
    Anyhow, you may be interested in :
    {message:id=9360007}
    {message:id=9984244}

  • How to create parameter and control file like filename + date

    Hello there
    I am trying to create parameter and control file with following command
    in SQLPLUS
    create pfile='/u03/oradata/WEBDB/backup/initWEBDB.ora' from spfile;
    In RMAN
    copy current controlfile to '/u03/oradata/WEBDB/backup/cf_longterm.cpy';
    how can I put date at the end of filename like
    initWEBDB8jan06.ora and cf_longterm8jan06.cpy
    Thanks in advance
    Lionel

    ASM is reliable but a smart DBA is very careful. If ASM is doing mirroring this is like RAID doing mirroring. What happens if you accidentally delete one copy ... the other one disappears instantly. Not a good idea.
    With respect to redo logs you need a minimum of three groups, two members, and one thread per instance. So a 2 node cluster should, at a minimum have 12 physical files.
    Not mirroring the redo logs, assuming multiple members, is not as critical.

  • The backup file contains unrecognized data and cannot be used

    I am trying to restore files from an NT back up on a SBS 2003.   The backup was taken on this machine and trying to restore to the same.   I get this error: The backup file contains unrecognized data and cannot be used.
    I tried copying it to a different machine with no luck.  Any suggestions?
    Robert

    Hi,
    Since the backup file cannot be restored on a different machine, the backup file could be corrupt. You need to use some third party recovery utilities to restore the backup file.
    For more detailed information, please refer to the article below:
    Windows NTBackup restore troubleshooting tips
    http://searchdatabackup.techtarget.com/tip/Windows-NTBackup-restore-troubleshooting-tips
    Please Note: Since the website is not hosted by Microsoft, the link may change without notice. Microsoft does not guarantee the accuracy of this information.
    Best Regards,
    Mandy 
    If you have any feedback on our support, please click
    here .
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

  • Spool SQl data into text file using dynamic sql

    Hi,
    I am spooling output data into text file using command
    select 'select t.mxname,bo.lxtype,t.mxrev'||chr(10)||'from mx_1234567'||chr(10)||
    'where <condition>';
    here mxname varchar(128),lxtype(128),mxrev(128) all are of varchar type.I want the output in format
    e.g Part|1211121313|A
    but due to column width the output,I am getting is with spaces.
    "Part then blank spaces |1211121313 then blank spaces |A"
    how can I remove these spaces between columns.I used set space 0 but not working.
    Thanks in advance.
    Your help will be appreciated.

    Hi Frank,
    I have seen your reply for SET LINE SIZE function. But, I could not be able to understand it.
    I am facing similar kind of issue in my present project.
    I am trying spool more than 50 columns from a table into flat file. Because of more column lengths in few columns, i am getting space. There are so many columns with the same issue. I want to remove that space.so that, data can fit perfectly in one line in .txt file without any wrap text.
    Below is my sample query.sql. Please let me know the syntax. My mail id : [email protected]
    --Created : Sep 22,2008, Created By : Srinivasa Bojja
    --Export all Fulfillments
    --Scheduled daily after 1:00am and should complete before 3:30am
    WHENEVER SQLERROR EXIT SQL.SQLCODE
    SET LINESIZE 800
    SET WRAP OFF
    SET PAGESIZE 800
    SET FEEDBACK OFF
    SET HEADING ON
    SET ECHO OFF
    SET CONCAT OFF
    SET COLSEP '|'
    SET UNDERLINE OFF
    SPOOL C:\Fulfillment.txt;
    SELECT SRV.COMM_METHOD_CD AS Method,
    SRV.SR_NUM AS "Fulfillment Row_Id",
    CON.LAST_NAME AS "Filled By"
    SRV.SR_TITLE AS Notes,
    SRVXM.ATTRIB_04 AS "Form Description"
    FROM SIEBEL.S_SRV_REQ SRV,
    SIEBEL.S_SRV_REQ_XM SRVXM,
    SIEBEL.S_USER USR,
    SIEBEL.S_CONTACT CON
    WHERE SRV.ROW_ID = SRVXM.PAR_ROW_ID AND
    SRV.OWNER_EMP_ID = USR.ROW_ID AND
    CON.ROW_ID= SRV.CST_CON_ID;
    SPOOL OFF;
    EXIT;

  • Change the Data and Log file locations in livecache

    Hi
    We have installed livecache in unix systems in the /sapdb mount directory where the installer have created sapdata and sapdblog directories. But the unix team has already created two mount direcotries as follows:
    /sapdb/LC1/lvcdata and /sapdb/LC1/lvclog mount points.
    While installing livecache we had selected this locations for creating the DATA and LOG volumes. Now they are asking to move the DATA and LOG volumes created in sapdata and saplog directories to these mount points. How to move the data and log file and make the database consistent. Is there any procedure to move the files to the mount point directories and change the pointers of livecahce to these locations.
    regards
    bala

    Hi Lars
    Thanks for the link. I will try it and let u know.
    But this is livecache (even it uses MaxDB ) database which was created by
    sapinst and morover is there any thing to be adjusted in SCM and as well as
    any modification ot be done in db level.
    regards
    bala

  • Export batch data into CSV file using SQL SP

    Hi,
    I have created WCF-Custom receive adapter to poll Sql SP (WITH xmlnamespaces(DEFAULT 'Namespace' and For XML PATH(''), Type) . Get the result properly in batch while polling and but getting error while converting into CSV by using map.
    Please can anyone give me some idea to export SQL data into CSV file using SP.

    How are you doing this.
    You would have got XML representation for the XML batch received from SQL
    You should have a flat-file schema representing the CSV file which you want to send.
    Map the received XML representation of data from SQL to flat-file schema
    have custom pipeline with flat-file assembler on the assembler stage of the send pipeline.
    In the send port use the map which convert received XML from SQL to flat file schema and use the above custom flat-file disassembler send port
    If this answers your question please mark it accordingly. If this post is helpful, please vote as helpful by clicking the upward arrow mark next to my reply.

  • Importing a .DAT and .001 files

    Hi,
    I need to create a flat file (ascii) from a .DAT and .001 files created by the Oracle Export utility.
    I thought that I could import the above mentionned fileds to Oracle 8i Lite and then export the tables to a flat file (ascii).
    So far I could not find a way to import those files with Lite 8i.
    I'm curently downloading the Data Migration utylity.
    What is the best way to proceed?
    Thank you.

    if you already have a dat file..then what seems to be the problem..
    I am confuse..however if you have the dat file..say from example..
    I have a dat file and the data in my dat look like the following:
    'Blaze',1,'Available', 123,
    and the control files look like the following:
    LOAD DATA
    APPEND
    INTO TABLE table_name
    FIELDS TERMINATED BY ","
    TRAILING NULLCOLS
    (PRIZE CHAR(256) OPTIONALLY ENCLOSED BY "'" ,
    PROMOTION_ID,
    STATUS CHAR(20) OPTIONALLY ENCLOSED BY "'" ,
    USER_ID)
    then all you need to do is use sql loader..
    good luck
    shah

  • Reading uknown DB *.dat and *.idx files

    I'm trying to read the tables structure and data of a couple of *.dat and *.idx files. I don't know what DB are they created with - I tried diffrent db managers but noone recognized them. Some file examples are here (rar archive 304 kb): http://zorko.ru/_db2read.rar
    Any help much appreciated.

    well, you can't really tell what's database these files are for just by extension of filename. Oracle will take any file name extension that does not violate the rule.
    the wild guess is .dat is datafiles of tablespace holding data and .idx is datafiles of tablespaces holding indexes.
    However if they do for Oracle database you can't read out meaning information from them without properly setup the database. To do that you will need more information.
    try to use "strings" to read the file on unix platform, see if you can get anything ring the bell.

  • Not able to extract performance data from .ETL file using xperf commands. getting error "Events were lost in this trace. Data may be unreliable ..."

    Not able to extract  performance data from .ETL file using xperf commands.
    Xperf Commands:
    xperf –i C:\TempFolder\Test.etl -o C:\TempFolder\BootData.csv  –a process
    Getting following error after executing above command:
    "33288636 Events were lost
    in this trace. 
    Data may be unreliable
    This is usually caused
    by insufficient disk bandwidth for ETW lo
    gging.
    Please try increasing the minimum
    and maximum number of buffers
    and/or
                    the buffer size. 
    Doubling these values would be a good first at
    tempt.
    Please note, though, that
    this action increases the amount of me
    mory
                    reserved
    for ETW buffers, increasing memory pressure on your sce
    nario.
    See "xperf -help start"
    for the associated command line options."
    I changed page size file but its does not work for me.
    Any one have idea, how to solve this problem and extract ETL file data.

    I want to mention one point here. I have total 4 machines out of these 3 machines above
    commands working properly. Only one machine has this problem.<o:p></o:p>
    Hi,
    I consider that you can try to use xperf to collect the trace etl file and see if it can be extracted on this computer:
    Refer to following articles:
    start
    http://msdn.microsoft.com/en-us/library/windows/hardware/hh162977.aspx
    Using Xperf to take a Trace (updated)
    http://blogs.msdn.com/b/pigscanfly/archive/2008/02/16/using-xperf-to-take-a-trace.aspx
    Kate Li
    TechNet Community Support

  • Display data in log file using PL/SQL procedure

    Just as srw.message is used in Oracle RDF Reports to display data in log file in Oracle Apps, similarly how it is possible to display data in log file using PL/SQL procedure?
    Please also mention the syntax too.

    Pl post details of OS, database and EBS versions.
    You will need to invoke the seeded FND_LOG procedure - see previous discussions on this topic
    Enable debug for pl/sql
    https://forums.oracle.com/forums/search.jspa?threadID=&q=FND_LOG&objID=c3&dateRange=all&userID=&numResults=15&rankBy=10001
    HTH
    Srini

  • I have Creative Cloud installed and Photoshop installed, but I can not open the program itself. No idea why, everything is up to date and i was using it just fine yesterday.

    I have Creative Cloud installed and Photoshop installed, but I can not open the program itself. No idea why, everything is up to date and I was using it just fine yesterday.
    2015-02-24_1129 - DavidIpromote's library

    Try resetting Photoshop Preferences.   Use your Photoshop start icon to launch Photoshop and immediately press and hole Shift+Ctrl|CMD+Alt|Option keys   Photoshop should Prompt you with Do you want your Preferences deleted. Reply yes.   Photoshop should then delete your user ID preferences and then create a default set of preferences for your ID and start successfully.

  • How to write data to text file using external tables

    can anybody tell how to write data to text file using external tables concept?

    Hi,
    Using external table u can load the data in your local table in database,
    then using your local db table and UTL_FILE pacakge u can wrrite data to text file
    external table
    ~~~~~~~~~~~
    http://download-east.oracle.com/docs/cd/B19306_01/server.102/b14200/statements_7002.htm#i2153251
    UTL_FILE
    ~~~~~~~~~
    http://download-east.oracle.com/docs/cd/B19306_01/appdev.102/b14258/u_file.htm#sthref14093
    Message was edited by:
    Nicloei W
    Message was edited by:
    Nicloei W

Maybe you are looking for

  • How to delete files from JAR ?

    Hi, Is there any way to delete a file from the jar ? I just want to delete some class files from the jar without deleting and again uploading the JAR. Thanks & Regards, Atur Shah

  • How to upload file on Application Server with Forms 6i?

    Please, I need to upload .csv file from local to Application Server, I think to use .jsp application, but I don't know where I had to put it...Apache can't read .jsp file without Tomcat? If I launch my .jsp file from Apache/htdocs directory can't vie

  • Interface map not exist in cache

    hi iam configuring proxy to bapi and getting following error Interface mapping urn:xiworkshop:group00:webapp PO_out__ZBAPI_PO_CREATE does not exist in runtime cache i refreshed the cache and done reactivation of the objects still same story. plz help

  • XMLBean generation from WSDL file

    I have used workshop for weblogic 9.2 to generate web service from wsdl files. The problem is however, that the (XmlBean) code generated from the <wsdl:fault name="..... /> does not genetate a "Throwable" object. The resulting methods produced by the

  • Starvation in MDBs ?

    Our application receives 10 different types of messages on one queue each. Thus we have 10 queues (MQ as a Foreign JMS provider with MDBs in a WLS). We have MDBs processing each of these queues. The producer (mainframe) that sends messages to these q