Help required to manipulate the data without exporting the output twice

I am running a script which export the data to CSV file and I re-import the same file to remove the duplicate rows and them export them.
Need to clean the up the first file after the script execution
Is it possible to store the data in an object without export and import for manipulating the data
&{foreach($role in Get-VIRole){
    Get-VIPrivilege -Role $role -ErrorAction SilentlyContinue |
    Select @{N="Role";E={$role.Name}},@{N="Assigned Privileges";E={$_.ID}}
}} | Export-Csv C:\Scripts\Permission.CSV -NoTypeInformation -UseCulture 
&{foreach($row in (Import-Csv C:\Scripts\Permission.CSV -UseCulture)){   
        if($prevRole -and $row.Role -eq $prevRole){
            $role = ""
        Else{
            $role = $prevRole = $row.Role
        New-Object PSObject -Property @{
            Role = $role
           "Assigned privileges" = $row."Assigned Privileges"
    }} | Select "Role","Assigned Privileges" | Export-CSV

If I understand the question correctly, I think this might work.
$ht = @{}
foreach($role in Get-VIRole){
Get-VIPrivilege -Role $role -ErrorAction SilentlyContinue |
foreach { $ht[$role.name] = $_.ID }
$ht.GetEnumerator() |
foreach {
New-Object PSObject -Property @{Role=$_.name;'Assigned Privileges' = $_.value}
} | export-csv
That uses a hash table to de-dupe the role/assignment pairs and then builds the objects for export from the hash table.
[string](0..33|%{[char][int](46+("686552495351636652556262185355647068516270555358646562655775 0645570").substring(($_*2),2))})-replace " "

Similar Messages

  • Help required to understand the concepts for data files in tablespace

    Hi Guys!
    Conceptual understanding for data files.
    A tablespace is having 2 datafiles.
    What once the data files will get full? I know we can create data files but what if there is no space on the server for any additional file to be created?
    Do we require to make the oldest file offline and move outside the environment?
    Thanks

    782790 wrote:
    Hi Guys!
    Conceptual understanding for data files.
    A tablespace is having 2 datafiles.
    What once the data files will get full? I know we can create data files but what if there is no space on the server for any additional file to be created?
    Do we require to make the oldest file offline and move outside the environment?
    ThanksIf the datafiles are full, there is not much that you would get from the "defragmentation" of the datafiles. It would be sort of like this in the start,
    FFFFFFFFFFFFFFFFFFFFFFFFFF
    Then it will be ,
    FFFFFFFUUUUUFFFFFUUUUUU
    So what you are going to get is something like this, (if would be done),
    FFFFFFFFFFFFFUUUUUUUUUU
    So that's how the datafile would like. But that won't give you much of the space anyways still. The option to add a storage must be there with you. The file being offline wont take away the space that it has already allocated. And moreover, doing so, would also impact the access to the data that it would be storing within it.
    HTH
    Aman....

  • Urgent Help required! - Storing the XML as String instead as a file

    Hi,
    I need urgent help on this.
    I have an XML file. I have used org.w3c.dom to build dom and manipulate the XML file.
    I have updated the values for some of the nodes and I have deleted some of the unwanted nodes.
    I am able to save the output of the DOM as another XML file using
    either transform class or XMLSerializer with OutputFormatter class.
    But my requirement is to save the output of the DOM into a String instead of a file.
    When I save it in String, I need to have the following XML decalration and DOCTYPE declration also with it.
    <?xml version="1.0" encoding="UTF-8"?>
    <!DOCTYPE Test SYSTEM "Test.dtd">
    Can anyone pls help me in this??
    Thanks in Advance. Expecting some inpputs pls....!
    Regards,
    Gayathri.

    hi,
    i think this is what u want
        public static String getXmlString(Document d) {
          StringWriter strResponse = null;
          try {
             OutputFormat format  = new OutputFormat(d);
             strResponse = new StringWriter();
             XMLSerializer serial = new XMLSerializer( strResponse, format );
             serial.asDOMSerializer();
             serial.serialize(d.getDocumentElement());
          catch (Exception e) {
            System.out.println(e.toString());
          return strResponse.toString();
    }HTH
    vasanth-ct

  • Help Required: Working with relational data

    Hi,
    I'm looking for some advice
    Scenario
    We have a customer who has an SE relational database with Apex for transaction processing. They require reporting functionality and are debating whether to buy BI SE One or BI EE; they will not buy the enteprise edition database though.
    Questions
    Here are my questions:
    1. In order to use answers and dashboards, does the relational data have to be in a dimensional format. Is this reccommended.
    2. We were debating on whether to create data marts for various business requirements. Is this standard procedure for relational data (Which is not too complex)
    3. If the data has to be in a dimensional format, is this best modelled using warehouse builder or the model layer of the repository using BI Administrator. (Baring in mind that the customer is an SME using a SE database)
    The main problem where having, been fairly new to warehouse building, is that we dont know what the best path is to take when working with relational data. Do we create dimensional models? What should we use? Is a relational structure suitable enough?
    Can anyone shed light on our problems
    Regards
    Kevin

    1. In order to use Answers and Dashboards you do not need to have a Data Warehouse. It can report out of both Transactional database as well as a data warehouse.
    2. It depends on what your business requirements are. BI EE/SE1 does not require you to have data marts. Data Marts would give performance gains, would help you in visualizing the transactional data from a business standpoint, is a big plus but not a mandate. If you want to go that route of creating Datamarts, you can easily do that using BI SE1 since that gives you warehouse builder which can be both used as a data modelling tool and a ETL tool.
    3. Warehouse Builder is a ETL tool. It helps in you visualizing/deploying data warehouses in databases. But BI EE Admin tool is more of a logical dimension modelling tool wherein the facts and dimensions do not exist physically but the tool uses them to do the querying/processing.
    Since you are just starting, what you can do is to use BI EE directly on top of the transactional system for the time being. But as a long term solution you can start working on your data marts and as soon as they are ready you can port the BI EE on top of them rather than the transactional system.
    Thanks,
    Venkat
    http://oraclebizint.wordpress.com

  • Help required while publishing product data in Adobe InDesign

    Hi Experts,
    I have installed MDM7.1 publisher(build version- 7.1.01.46) and Adobe InDesign CS3(version-5.0.4) in my local system.
    I also installed the plug-in for InDesign version CS3 - the Microsoft Visual C++ 2005
    SP1 Redistributable Package (x86) in my local system.
    While publishing data from MDM publisher to Adobe InDesign, it's not populating data in Adobe InDesign window.
    What would be the problem while publishing, is there any other plu-ins required apart from the above.
    Please help me if any have idea's related this.
    Thanks,
    RDNPrasad

    Hi Mark
    Yes Indesign CS3 was running
    Instead of  I entered my computer name at the network, and after some tries it worked to some extent
    We got it working, but at the end of every publication,  Indesign stopped with an error message and was shut down.
    I tried to publish again today, but now nothing happens, no page is generated in Indesign
    I tried to generate a XML file and got this "IE7 error" on a 0 kb file:
    Cannot view XML input using style sheet. Please correct the error and then click the Refresh button, or try again later.
    The plugin is: MDM Paper Module, Version 7.1.02.59
    Indesign CS3, version 5.0.4
    MDM publisher : 7.1.01.46
    Kind regards
    Kurt
    Edited by: Kurt Jørgensen on May 27, 2009 12:02 PM
    Edited by: Kurt Jørgensen on May 27, 2009 12:03 PM

  • Help required in understanding the flow of control in working with EJB 3.0

    *1. Can anyone help me in understanding the flow after getting an object of InitialContext in using Stateless/Stateful in EJB 3.0 from client i.e. what we are getting in the step Object ref=jndiContext.lookup("CalculateBean/Remote) .*
    *2. What is the use of INITIAL_CONTEXT_FACTORY*
    *3. PROVIDER_URL*
    *4. Java.naimg.factory.url.pkgs*
    *5. why org.jboss.naming and org.jnp.interfaces separated by semi colon.*
    *6. Why we dont require INITIAL_CONTEXT_FACTORY, PROVIDER_URL and Java.naming.factory.url.pkgs if the client is a web client.*
    Properties p=new Properties();
              p.put(Context.INITIAL_CONTEXT_FACTORY, "org.jnp.interfaces.NamingContextFactory");
              p.put(Context.PROVIDER_URL,"jnp://localhost:1099");
              p.put("java.naming.factory.url.pkgs","org.jboss.naming:org.jnp.interfaces");
              Context jndiContext = new InitialContext(p);
         Object ref = jndiContext.lookup("CalculateBean/Remote");
    Calculate c = (Calculate)ref;
    int sum=c.add(3,4);
    System.out.println(sum);
    Thanx in advance........

    Annotations can also be processed at run time using the core reflection API (java.lang.Class & java.lang.reflect.*), so you might want to explore down that path. I have no idea how JBoss actually does it, but annotation processing at build time and core reflection at runtime are the two most obvious possibilities.
    Bruce

  • Help required in implementing the algorithm

    Actually I am brand new to the java programming, and i need some help to implement the algorithm below in java.
    I have the dataset required for the execution,
    Can any body please help me?
    Input: (1) a source database D,
    (2) a min support,
    (3) a min confidence,
    (4) a set of recommended items Y
    Output: a transformed database D, where rules containing
    Y on RHS will be hidden
    1. Find large 1-item sets from D;
    2. For each hidden item y �� Y
    3. If y is not a large 1-itemset, then Y := Y &#8722; {y};
    4. If Y is empty, then EXIT;// no AR contains Y in RHS
    5. Find large 2-itemsets containing items in Y from D;
    6. For each y �� Y {
    7. For each large 2-itemset containing y {
    8. Compute confidence of rule U, where U is a rule likex �� y;
    9. If confidence(U) �� min conf, then { //Calculate # of transactions required to lower the confidence of U
    10. Calculate iterNumConf = _|D|&#8727;(supp(xy) &#8722; minconf &#8727;supp(x))_;
    //Calculate # of transactions required to lower the support of U
    11. Calcualte iterNumSupp = _|D|&#8727;(supp(xy)&#8722; min supp)_;
    12. Calculate k = minimum (iterNumConf, iterNumSupp);
    13. Calculate |TR| = |xy|; //|TR|: # of transactions inTR
    14. if k > |TR|
    15. Can not hide U;
    16. else {
    17. Find T kR = first k transactions that contain xy;
    18. Sort T kR in ascending order by the number of items;
    19. for i = 1 to iterNum do { 20. Choose the first transaction t from T kR ;
    21. Modify t so that y is not supported, the RHS(U);
    22. |xy| = |xy| &#8722; 1; // update support count of xy
    23. confidence(U) = |xy|/|x|; // update confidence
    24. Remove and save the first transaction t from T kR ;
    25. }; // end for iterNum
    26. }; // end if confidence(U)
    27. }; // end of each large 2-itemset
    28. Remove y from Y;
    29. } // end of for each y
    30. Output updated D, as the transformed D;

    The way you post that, it looks very much like "Do my work for me". Nobody around here likes that kind of post, they consider it lazy and selfish. But perhaps you didn't mean it that way? Why don't you explain what you have done and what parts of it you are having problems with. Explain those problems in detail.

  • Help required "Forming Packets od data "

    Hello
    i need help in forming packets from data which is read from aport and transfer those packets to the another server.
    can any one one help me ,which java class can be used for this.

    (crappy forum)
    Double post:
    http://forum.java.sun.com/thread.jspa?threadID=728704

  • Help required regarding tunning the query mentioned

    HI all ,
    Query mentioned below takes around 1 hr to complete . It's being used by the autoconfig kindly me in tunning it ..
    QUery :
    UPDATE WF_ITEM_ATTRIBUTE_VALUES WIAV SET WIAV.TEXT_VALUE = REPLACE(WIAV.TEXT_VALUE,:B1,:B2)
    WHERE (WIAV.ITEM_TYPE, WIAV.NAME) = (SELECT WIA.ITEM_TYPE, WIA.NAME
    FROM WF_ITEM_ATTRIBUTES WIA WHERE WIA.TYPE = 'URL'
    AND WIA.ITEM_TYPE = WIAV.ITEM_TYPE
    AND WIA.NAME = WIAV.NAME)
    AND WIAV.TEXT_VALUE IS NOT NULL
    AND INSTR(WIAV.TEXT_VALUE
    , :B1) > 0
    Plan :*
    <pre>
    | Id | Operation | Name | Rows | Bytes | Cost |
    | 0 | UPDATE STATEMENT | | 453 | 14496 | 284K|
    | 1 | UPDATE | WF_ITEM_ATTRIBUTE_VALUES | | | |
    |* 2 | FILTER | | | | |
    |* 3 | TABLE ACCESS FULL | WF_ITEM_ATTRIBUTE_VALUES | 453 | 14496 | 282K|
    |* 4 | TABLE ACCESS BY INDEX ROWID| WF_ITEM_ATTRIBUTES | 1 | 33 | 2 |
    |* 5 | INDEX UNIQUE SCAN | WF_ITEM_ATTRIBUTES_PK | 1 | | 1 |
    Predicate Information (identified by operation id):
    2 - filter(("SYS_ALIAS_2"."ITEM_TYPE","SYS_ALIAS_2"."NAME")= (SELECT /*+ */
    "WIA"."ITEM_TYPE","WIA"."NAME" FROM "APPLSYS"."WF_ITEM_ATTRIBUTES" "WIA" WHERE
    "WIA"."NAME"=:B1 AND "WIA"."ITEM_TYPE"=:B2 AND "WIA"."TYPE"='URL'))
    3 - filter("SYS_ALIAS_2"."TEXT_VALUE" IS NOT NULL AND
    INSTR("SYS_ALIAS_2"."TEXT_VALUE",:Z)>0)
    4 - filter("WIA"."TYPE"='URL')
    5 - access("WIA"."ITEM_TYPE"=:B1 AND "WIA"."NAME"=:B2)
    </pre>
    Index :*
    <pre>
    INDEX_NAME COLUMN_NAME
    APPLSYS WF_ITEM_ATTRIBUTE_VALUES_PK 1 ITEM_TYPE
    2 ITEM_KEY
    3 NAME
    </pre>
    regds
    Rahul
    Edited by: RahulG on Jan 2, 2009 10:47 PM
    Edited by: RahulG on Jan 2, 2009 10:48 PM

    RahulG wrote:
    HI all ,
    Query mentioned below takes around 1 hr to complete . It's being used by the autoconfig kindly me in tunning it ..
    A few notes:
    1. Your query is using bind variables. If you're already on 9i or later (probably 9iR2 according to plan output), this statement will be subject to bind variable peeking and therefore the output of EXPLAIN PLAN is only of limited use, since the actual execution plan might be different and/or might be based on different cardinality estimates based on the actual bind values peeked at hard parse time. You can use the V$SQL_PLAN view to get the actual execution plan(s) if the statement is still cached in the shared pool, from 10g on DBMS_XPLAN.DISPLAY_CURSOR is available for that purpose.
    2. The execution plan posted suggests that only 453 rows will correspond to the filter criteria (but, as mentioned in 1. is based on an unknown bind variable value when using EXPLAIN PLAN), and probably therefore the optimizer didn't unnest the subquery but runs this as recursive FILTER query potentially for each row passing the filter criteria on the driving table WF_ITEM_ATTRIBUTE_VALUES. Depending on the actual number of rows this might be inefficient, and unnesting the subquery and turning it into a join might be more appropriate. This might accomplished e.g. by providing more representative statistics to the optimizer (are the statistics up-to-date?).
    Although you can't change the SQL you could try this manually by using the UNNEST hint to see if it makes any difference in the execution plan (and run time):
    WHERE (WIAV.ITEM_TYPE, WIAV.NAME) = (SELECT /*+ UNNEST */ WIA.ITEM_TYPE, WIA.NAME
    ...3. The composite index WF_ITEM_ATTRIBUTE_VALUES_PK can only be used on the first column ITEM_TYPE for effective index access, the NAME column would have to be used as filter on all index leaf blocks that would be found using a range scan on ITEM_TYPE. This might be quite inefficient, and/or might lead to a lot of rows/blocks that need to be visited in the table using this index access path.
    4. You could try to trace the execution by enabling extended SQL trace, e.g. using the (undocumented) DBMS_SUPPORT package in 9i. Running the "tkprof" utility on the generated trace file tells you the actual row source cardinalities (which can then be compared to the estimates of the optimizer) and - if the "waits" have been enabled - what your statement has waited for most.
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • Help Required in Migrating the Dev to Test Instance

    Hi
    Am very new to SOA platform, i developed one project in dev instance. now i would like to move this project to test instance.
    can any one help me to migrating the project.
    So I Need sample document with step by step instruction. i have 5 days only to migrate, i gone through so many blogs and post in metalink, but i didn't understand the concept exactly.
    Santhosh Kumar. K

    Hi K,
    There are many ways to do it, it will all depend on your environment.
    The way that works for me is:
    1. Deploy your project to SAR (it will actually generate a jar file)
    2. Copy the generated file to the environment where it will be deployed
    3. Use the ant-sca-deploy.xml to deploy it
    ant-sca-deploy.xml will be located at $BEA_HOME/Oracle_SOA1/bin
    and you can call it from your own ant build.xml like this:
         <target name="deploy-app">
              <ant
                   antfile="${beaHome}/Oracle_SOA1/bin/ant-sca-deploy.xml"
                   target="deploy"
                   dir="${beaHome}/Oracle_SOA1/bin"
                   >
                   <property name="serverURL" value="${scaDeployUrl}"/>
                   <property name="sarLocation" value="${appEarDestDir}/${sca.jar}"/>
                   <property name="overwrite" value="true"/>
                   <property name="user" value="${domainUser}"/>
                   <property name="password" value="${domainPwd}"/>
                   <property name="failOnError" value="true"/>
                   <property name="partition" value="${partition}"/>
              </ant>
         </target>
    Cheers,
    Vlad
    Give points - it is good etiquette to reward an answerer points (5 - helpful; 10 - correct) for their post if they answer your question. If you think this is helpful, please consider giving points

  • Help required to recognize the correct language import package

    Hi friends,
    I am trying to import ukranian(UK) language to SAP system using SMLT transaction.Please help me to identify the correct 'language import package'
    from the DVD dump.I cannot recognize the correct DVD .I am not clear how do I realize the exact package from the DVD dump.Kindly help me.
    Regards,
    Syam krishnan

    Hi Syam
    Please refer the TWO notes for clarity
    Note 97476 - Language availability R/3 4.0x - 4.7x, ECC500
    Note 517310 - Ukrainian language import has been filled w/ Russian
    Cheers,
    Shyam

  • SELECT stmt required to retrieve the output data as specified.

    source table:
    Origin      Destination      Miles
    Sydney           Melbourne      1000
    Perth           Adelaide      3000
    Canberra      Melbounre      700
    Melbourne      Sydney      1000
    Brisbane      Sydney      1000
    Perth      Darwin      4000
    Sydney      Brisbane      1000
    Duplicate routes are to be ignored so the output is
    out put:
    Origin           Destination      Miles
    Sydney           Melbourne      1000
    Perth           Adelaide      3000
    Canberra      Melbounre      700
    Brisbane      Sydney           1000
    Perth           Darwin           4000
    Please suggest a select query / sub query with out using any subprograms or functions/pkgs to get the out put table.

    Hi Pawan,
    Try this document
    <b>https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f0de1eb8-0b98-2910-7996-8a3c2fcf6785</b>
    Regards,
    ABY

  • Requirement to modify the output layout for F.19 transaction.

    Hi Friends,
      I need to modify the list output and need to add two more fields in the output. I have gone through the report and noticed that the field catalogue is being prepared using REUSEFIELDCAT*MERGE.  I have changed the internal table structure (LIST_ITEM) used to get the fld cat. But its not returning the added fields in the fld.cat. So please let me know if this can be done in any other way or at customization level.  Hope there is some thing with LIST_ITEM structure/name.
    Thanks & Regards,
    Venkat.

    Hi Friends,
      I need to modify the list output and need to add two more fields in the output. I have gone through the report and noticed that the field catalogue is being prepared using REUSEFIELDCAT*MERGE.  I have changed the internal table structure (LIST_ITEM) used to get the fld cat. But its not returning the added fields in the fld.cat. So please let me know if this can be done in any other way or at customization level.  Hope there is some thing with LIST_ITEM structure/name.
    Thanks & Regards,
    Venkat.

  • Urgent HELP required on forming the Matrix of data using PL/SQL

    Hi All,
    I'm new to this thread and require your urgent help in this regard.
    I've got a requirement for building a 5000 X 5000 matrix using PL/SQL. My original data tables have 5000 rows each and I need to do a correlation analysis using this data and need to store in a physical table and not in-memory. Is this feat achievable using mere PL/SQL? I understand that Oracle DB has a limitation of 1000 columns(but not sure) and hence I'd like to know whether there is any work-around for such scenarios. If not, what are the other alternative method(s) to achieve this feat? Do I need to use any 3rd party tools to get this done? An early reply from the experts is highly appreciated.
    Thanking you all Gurus in advance.
    Rgds
    Sai

    Welcome to OTN!
    I'll get to your quesiton in a moment, but first some welcome information. Many OTN posters consider it impolite to mark threads as "urgent". We are volunteers and have jobs of our own to do without people we don't know making demands. You are brand new and deserve some patience but please understand this. It is very likely before I finish this post someone will complain about the word "urgent" in your subject.
    On to more interesting things :)
    You can do the matrix, but are out of luck with a 5000 x 5000 table because Oracle only allows 1000 columns per table. There are ways to work around this.
    How do do the matrix depends on what you want to do. You can do this different ways. You can create a table beforehand and use PL/SQL or simple SQL to populate it, or use the CREATE TABLE AS syntax to create and populate it in one step if you can get the underlying SQL to work the way you want, something like
    create table my_table as
      select a.*, b.*
        from table1 a, table2 bcan populate a matrix from 2 tables with an intentional cartesian join (the WHERE clause was left out intentionally, provided your data is already in the data base.
    If not you can use a PL/SQL routine to populate the data.
    There are a couple of ways to solve the 1000 column limit. The easiest way might be to have 5 collections of 1000 columns each. A more complicated but more elegant soltion would be to have nested collections, allowing 2 colliections that you can loop through - a collection of collections. Nested collections can be hard to work with. A third way would be to use nested tables in the database but I personally do not like them and the insert, update, and delete statements for nested tables are hard to use.
    I'm not going to give a code example because I am not sure which solution is best for you. If you have further questions post them.

  • Urgent help required to write the code in  update routine

    Hi all,
    i want to calculate open purchase order qty in update routine
    formula is
    open purchase order qty = scl qty - rec qty
    where schd line date is less then or equal to 90 days from the current date.
    I have written one code : but its giving error that comm_structure is not defined in abap dictionary,can any body help to write appropriate routine. this calculation i am making for MM, and data source and cube are 2lis_02_scl and cube is zc_pur01 ( made by coping the 0pur_c01), data source scl has all fields required in the foumula...here is my code
    DATA:  COMM_STRUCTURE LIKE  /BIC/CS2LIS_02_SCL.
    DATA: SCL_QTY LIKE COMM_STRUCTURE-/BIC/ZK_SCLQTY.
    DATA:      GR_QTY LIKE COMM_STRUCTURE -/BIC/ZK_GRQTY.
    DATA: SCL_DATE.
    SCL_DATE = SY-DATUM + 90.
    SELECT COMM_STRUCTURE -/BIC/ ZK_SCLQTY COMM_STRUCTURE -/BIC/ ZK_GRQTY FROM
    /BIC/CS2LIS_02_SCL INTO   SCL_QTY     GR_QTY  
    WHERE
    COMM_STRUCTURE -/0SCHED_DATE LE SCL_DATE.
    IF SYSUBRC  = 0
    RESULT = SCL_QTY - GR_QTY.
    ELSE = NOVALUE.
    ENDIF.
    Can any body help me soon its very urgent.
    thanks

    Hi Anupam,
    I am not a very good ABAP Programmer, but found some things to notify u inyour code..
    DATA: COMM_STRUCTURE LIKE /BIC/CS2LIS_02_SCL,
          SCL_QTY LIKE COMM_STRUCTURE-/BIC/ZK_SCLQTY,
          GR_QTY LIKE COMM_STRUCTURE -/BIC/ZK_GRQTY.
    DATA  SCL_DATE LIKE SY-DATUM.
    SCL_DATE = SY-DATUM + 90.
    SELECT COMM_STRUCTURE -/BIC/ ZK_SCLQTY COMM_STRUCTURE -/BIC/ ZK_GRQTY FROM
    /BIC/CS2LIS_02_SCL INTO SCL_QTY GR_QTY
    WHERE
    COMM_STRUCTURE -/0SCHED_DATE LE SCL_DATE.
    IF SYSUBRC = 0
    RESULT = SCL_QTY - GR_QTY.
    ELSE.
    What should be the return value if you dont have to    calculate the result."
    ENDIF.
    try it.. and see..
    regards,
    kishore.

Maybe you are looking for

  • Warning for hierarchy in report

    Hi, when i execute a report with rsrt I get the following warning and no result at all... To the InfoObject Balance/P&L-Position ist the Hierarchy "" not available When I check the query in BEx Designer, I get the following error... To the InfoObject

  • What accessories do I need with my 3D Plasma TV??

    Do I need to purchase extra cables or features to get the full quality of the TV?  I have a blueray player and Apple TV.  Also does the TV come with a stand? or does that need to be purchased seperate too?

  • How can I set object as a parameter of one function

    I want to pass a object as one of the parameters so that I can execute the function of the object in another class. Urgent and thanks a lot.

  • JOptionPane Confusion!

    Hi, Im having a real difficult time with JOptionPanes! Im still a newbie to Swing and I just cant seem to get my head around how to make the JOptionPanes do what you want them to! I've read the tutorials and manuals but im still stuck. Basically, I h

  • Aperture 3 and web journal

    I'm trying to post a web journal to my mac account with little success. I'm using the Export Web pages button with the destination being Idisc.sites.journalname. The exporting ball spins at the bottom of the page .... for hours but never completes. I