Finding a common Time Dimension in DSO and InfoCube

Hi,
I am working on creating a MultiProvider which is a combination of 2 COPA InfoCubes, Open Orders DSO and a Shipment Data DSO.
I want to use the 'Requested Delivery Period' from Open Orders DSO and a custom field ZSHIPMONTH from the Shipments Data DSO for a common Period dimension in the MultiProvider but that does not look like an option when I tried to assign the 0FISCPER in the MultiProvider to these fields in the DSO. I do not see the fields from the DSOs as options for the period field here, I only see the 0FISCPER from the two COPA InfoCubes as seen in the screenshot attached.
The only alternative I can think of is to create separate InfoCubes on top of these two DSOs and then use a MultiProvider.
I learned that reporting is allowed on top of DSOs from 7.3 onwards, is there any easier way to get a common time dimension across these InfoCubes and DSOs for a MultiProvider?

Hi,
in both DSO and Cube the data is stored as
          Sales Doc1   M1  10   100
          Sales Doc1   M2    5     50
          Sales Doc2   M1  20   200.
The material M1 in sales doc1 does not get overwritten with M1 in sales doc 2  in DSO since the data is updated based on sales doc no.
If u try to view the contents of data target or analyse the data at reporting level with all these fields then it displays as above.
But if u ignore sales document field then the data is displayed as,
                                   M1  20   300
                                   M2   5      50.
In Cube, suppose if you dont have sales doc, then
Material | Qty | Amount
M1 30 300
M2 5 50
and in DSO.
Material | Qty | Amount
M1 20 100
M2 5 50
Thanks,
Sandeep

Similar Messages

  • What is the difference between DSo and Infocube

    Hello,
             Kindly tell me what is the difference between DSO and Infocube?
    And please tell me how to take the desicion that in whichi case we can use DSO and in which case we can use Infocube..

    Hi ,
    DataStore object serves as a storage location for consolidated and cleansed transaction data or master data on a document (atomic) level.
    This data can be evaluated using a BEx query.
    A DataStore object contains key fields (for example, document number/item) and data fields that can also contain character fields (for example, order status, customer) as key figures. The data from a DataStore object can be updated with a delta update into InfoCubes and/or other DataStore objects or master data tables (attributes or texts) in the same system or across different systems.
    Unlike multidimensional data storage using InfoCubes, the data in DataStore objects is stored in transparent, flat database tables. The system does not create fact tables or dimension tables.
    Use
    The cumulative update of key figures is supported for DataStore objects, just as it is with InfoCubes, but with DataStore objects it is also possible to overwrite data fields. This is particularly important with document-related structures. If documents are changed in the source system, these changes include both numeric fields, such as the order quantity, and non-numeric fields, such as the ship-to party, status and delivery date. To reproduce these changes in the DataStore objects in the BI system, you have to overwrite the relevant fields in the DataStore objects and set them to the current value. Furthermore, you can use an overwrite and the existing change log to render a source delta enabled. This means that the delta that is further updated to the InfoCubes, for example, is calculated from two successive after-images.
    An InfoCube describes (from an analysis point of view) a self-contained dataset, for example, for a business-orientated area. You analyze this dataset in a BEx query.
    An InfoCube is a set of relational tables arranged according to the star schema: A large fact table in the middle surrounded by several dimension tables.
    Use
    InfoCubes are filled with data from one or more InfoSources or other InfoProviders. They are available as InfoProviders for analysis and reporting purposes.
    Structure
    The data is stored physically in an InfoCube. It consists of a number of InfoObjects that are filled with data from staging. It has the structure of a star schema.
    The real-time characteristic can be assigned to an InfoCube. Real-time InfoCubes are used differently to standard InfoCubes.
    ODS versus Info-cubes in a typical project scenario
    ODS
    why we use ods?
    why is psa  & ods nessasary
    Hope this helps,
    Regards,
    CSM Reddy

  • How to create process chain with dso and infocube at a time

    hi friends,
      i have a FLATFILE in application server and i need to upload it to STANDARD DSO and i created a INFO PACKAGE and DTP for DSO.
    i have a STANDARD INFOCUBE to upload data from above DSO. here DSO is the source.  i want to schedule  this process in process chain. how could i do this. any one please give me solution.
         can any one explain me the options in RSPC and how they work and where we exactly use. this is very helpful to me. please give me guidance. thank you.
    thanks,
    sree

    First make sure you have already created the infopackage to load the data from server to the DSO and the DTP to load data from DSO to the Cube.
    Now goto Tcode RSPC.
    Choose Create process Chain option. provide appropriate technical name and description.
    Drag "Execute infopackage" process type from left hand side pane in RSPC.  Choose the infopackage to load the data from Server to DSO. you can find this infopackage using the F4 help in the above selected process type.
    Drag the "data transfer process" process type to load data from the Flat file data source above to the DSO.
    Then Drag the "Activate DataStore Request"  process type and select the DSO in which you would load the data  using the above DTP.
    Drag the "data transfer process" process type to load data from DSO to the Cube.
    Save, activate  and schedule the process chain according to appropriate time.

  • Time Dimension 2 PARENTH2 and PRIOR,NEXT

    Hi all,
    We have standart TIME hierarchy from AppShell (MONTH, QUARTER, YEAR). But in some cases users want insert data in detalization by week. We decide add additional ID's with detalizations to weeks and add they in second hierarchy. Everything is ok in system.
    But than we try write scrip logic with  PRIOR or NEXT which use data from previous year it doesn't work. But than we use PRIOR or NEXT with current year months everything ok.
    Example of logic.
    *XDIM_MEMBERSET TIME = PRIOR, %TIME_SET%, %YEAR%.DEC
    *CALC_EACH_PERIOD
    *WHEN P_ACCT
    *IS A01
    *REC(EXPRESSION=%VALUE% - GET(P_ACCT="A02",TIME=PRIOR), P_ACCT = A03)
    *ENDWHEN
    *ENDWHEN
    Then we save data on January it doesn't work. BUt on December prev year A02 and January A01 we have data.
    Maybe you have some advises according second TIME dimension hierarchy. I think problem on it.
    Best regards,
    iaroslav

    Hi,
    Here is fragment of debuglogic.
    Start time --->7:29:17 PM  -  Date:9/28/2011  (build code:7.5.107)
    User:T2RU\bpc.sysadmin
    Appset:DEVTELE2
    App:BUDGETS
    Logic mode:0
    Logic by:
    Scope by:CATEGORY,INPUTCURRENCY,P_BF,P_BUDDIM,P_BUDDIM2,P_BUSTYPE,P_CC,P_CITY,P_DATASRC,P_PROJECT,TIME
    Data File:
    Debug File:D:\PC_MS\DATA\WebFolders\DEVTELE2\BUDGETS\PrivatePublications\bpc.sysadmin\TempFiles\DebugLogic.log
    Logic File:
    Selection:DIMENSION:CATEGORY|BU|DIMENSION:INPUTCURRENCY|RUR|DIMENSION:P_BF|IBF19|DIMENSION:P_BUDDIM|NO_BUDDIM|DIMENSION:P_BUDDIM2|NO_BUDDIM2|DIMENSION:P_BUSTYPE|NO_PRODUCT|DIMENSION:P_CC|3_KEM_FF_FA|DIMENSION:P_CITY|KEM|DIMENSION:P_DATASRC|CCDATA|DIMENSION:P_PROJECT|PR_PL|DIMENSION:TIME|2012.JAN|
    Run mode:1
    Query size:0
    Delim:,
    Query type:0
    Simulation:0
    Calc diff.:1
    Formula script:
    Max Members:
    Test mode:0
    Is Modelling:1
    Query Type:0
    Max members:
    Region:
    DIMENSION:CATEGORY
    BU
    DIMENSION:INPUTCURRENCY
    RUR
    DIMENSION:P_ACCT
    ----- There a lot of accounts <All> accounts and I dont show it all
    DIMENSION:P_BF
    IBF19
    DIMENSION:P_BUDDIM
    NO_BUDDIM
    DIMENSION:P_BUDDIM2
    NO_BUDDIM2
    DIMENSION:P_BUSTYPE
    NO_PRODUCT
    DIMENSION:P_CC
    3_KEM_FF_FA
    DIMENSION:P_CITY
    KEM
    DIMENSION:P_DATASRC
    CCDATA
    DIMENSION:P_PROJECT
    PR_PL
    DIMENSION:TIME
    2012.JAN
    DIMENSION:P_CC
    3_KIR_FF_FA
    DIMENSION:CATEGORY
    BU
    DIMENSION:P_DATASRC
    CCDATA
    DIMENSION:INPUTCURRENCY
    RUR
    DIMENSION:P_BF
    IBF19, IBF15
    DIMENSION:P_BUDDIM
    NO_BUDDIM
    DIMENSION:P_BUSTYPE
    NO_PRODUCT
    DIMENSION:P_BUDDIM2
    NO_BUDDIM2
    DIMENSION:TIME
    PRIOR, %YEAR%.JAN, %YEAR%.FEB, %YEAR%.MAR, %YEAR%.APR, %YEAR%.MAY, %YEAR%.JUN, %YEAR%.JUL, %YEAR%.AUG, %YEAR%.SEP, %YEAR%.OCT, %YEAR%.NOV, %YEAR%.DEC
    DIMENSION:P_ACCT
    <ALL>
    Loading TIME.TIMEID
    Time to load properties:0.0 sec.
    select P_ACCT,P_BF,TIMEID,SIGNEDDATA
    into #tblTempLogic_266166
    from tblFactBUDGETS
    WHERE  CATEGORY=N'BU' AND  INPUTCURRENCY=N'RUR' AND  P_BF in (N'IBF19',N'IBF15') AND  P_BUDDIM=N'NO_BUDDIM' AND  P_BUDDIM2=N'NO_BUDDIM2' AND  P_BUSTYPE=N'NO_PRODUCT' AND  P_CC=N'3_KIR_FF_FA' AND  P_CITY=N'KIR' AND  P_DATASRC=N'CCDATA' AND  P_PROJECT=N'PR_PL' AND  TIMEID in (N'20120400',N'20120800',N'20121200',N'10000425',N'20120200',N'20120100',N'20120700',N'20120600',N'20120300',N'20120500',N'20121100',N'20121000',N'20120900')
    insert into #tblTempLogic_266166 (P_ACCT,P_BF,TIMEID,SIGNEDDATA)
    select P_ACCT,P_BF,TIMEID,SIGNEDDATA
    from tblFactWBBUDGETS
    WHERE  CATEGORY=N'BU' AND  INPUTCURRENCY=N'RUR' AND  P_BF in (N'IBF19',N'IBF15') AND  P_BUDDIM=N'NO_BUDDIM' AND  P_BUDDIM2=N'NO_BUDDIM2' AND  P_BUSTYPE=N'NO_PRODUCT' AND  P_CC=N'3_KIR_FF_FA' AND  P_CITY=N'KIR' AND  P_DATASRC=N'CCDATA' AND  P_PROJECT=N'PR_PL' AND  TIMEID in (N'20120400',N'20120800',N'20121200',N'10000425',N'20120200',N'20120100',N'20120700',N'20120600',N'20120300',N'20120500',N'20121100',N'20121000',N'20120900')
    and SOURCE = 0
    insert into #tblTempLogic_266166 (P_ACCT,P_BF,TIMEID,SIGNEDDATA)
    select P_ACCT,P_BF,TIMEID,SIGNEDDATA
    from tblFAC2BUDGETS
    WHERE  CATEGORY=N'BU' AND  INPUTCURRENCY=N'RUR' AND  P_BF in (N'IBF19',N'IBF15') AND  P_BUDDIM=N'NO_BUDDIM' AND  P_BUDDIM2=N'NO_BUDDIM2' AND  P_BUSTYPE=N'NO_PRODUCT' AND  P_CC=N'3_KIR_FF_FA' AND  P_CITY=N'KIR' AND  P_DATASRC=N'CCDATA' AND  P_PROJECT=N'PR_PL' AND  TIMEID in (N'20120400',N'20120800',N'20121200',N'10000425',N'20120200',N'20120100',N'20120700',N'20120600',N'20120300',N'20120500',N'20121100',N'20121000',N'20120900')
    select tmpTable.P_ACCT,tmpTable.P_BF,tmpTable.TIMEID,sum(SIGNEDDATA) as SIGNEDDATA
    from #tblTempLogic_266166 as tmpTable
    group by tmpTable.P_ACCT,tmpTable.P_BF,tmpTable.TIMEID
    drop table #tblTempLogic_266166
    I saw that TIMEID 10000425 is incorrect -its 2012.DEC.WEEK3 from second hierarchy
    Thanks,
    Bets regards
    Iaroslav

  • How to Map the Unit field  in case of DSO and INFOCUBE

    Dear Experts,
    I have a issue ,Please help me to solve this
    I have DSO as provider ,
    And, i have to map transformations  btw the Datasource and DSO.
    In generic Data source,  i have unit fields like BASME,MEINS (Quantity units) & STWAE (currency field)
    and normal Quantity fields  like KWMNG,OAUME(quantity related),OAUWE (value related).
    In DSO data fields as Key figure info objects like  0Quantity (which have 0Unit as unit of measure) and some other  key figures which have there respective unit of measure in info object  definition.
    So you Please tell me how to map the Quantity ,Amounts, unit fields to key figures that we have.
    (How it will be for both DSO and Info cube is there any difference?)
    Edited by: AnjunathNaidu on Jan 18, 2012 1:20 PM

    Navasamol ,
    If it is works ,will u please tell me what is the difference ,if the transformations btw data source and DSO and
    what is the difference btw data source and info cube and btw DSO to Infocube or cube to cube .
    And i have  seen the Quantity fields  and there respective unit fields are mapped directly  to key figure info object
    in case of Info cube . Its working fine .
    If only 1:1 mapping allowed in DSO data fields key figures and there respective unit of measure characteristic.
    why this difference btw DSO and Info cube can any one explain me in detail.
    Expecting your valuable suggestions.
    Thanks & Regards,
    Anjunath Naidu
    Edited by: AnjunathNaidu on Jan 18, 2012 4:05 PM

  • Handling Time Dimensions with Monthly and Weekly analysis

    I'm charged to build a cube that can handle both monthly and weekly analysis. They want to be view projections on a weekly grain, but looking back at history they want to be able to tie back to the general ledger at a monthly level.
    Is it overkill to load data down to the daily level in order to support multiple roll ups to the month and week? What is the best way to handle this kind of situation?

    Having the data come as two different formats (one for months, one for weeks) versus having it come in daily buckets and constructing the months/weeks within Essbase is a call I wouldn't make without understanding the needs first.
    Assuming there is no need for data at the daily level, I would weigh the selection based on performace. If the performance is not too outragious, taking the data in at a daily basis would make it easier to reconstruct later. If performance is an issue (or becomes an issue), I would still be tempted to load a daily cube at level 0, summarize the Months and Weeks, and then feed the main cube from there.
    So many ways to do this... and for every way there is a reason you might want to try something else. I'd start with the easy approach, and only go "grand" if you have to (meaning: load it daily, summarize in Essbase, and revisit the detail/summary cube split idea if the performance tanks out).

  • Delta Load on DSO and Infocube

    Hi All,
            I would like to know the procedure for the scenario mentioned below.
    Example Scenario:
    I have created a DSO with 10 characteristics and 3 keyfigure. i have got 10 Customers whose transactions are happening everyday. A full upload on the DSO was done on 7th October 09. How can i load their changing data's from 8th Oct to till date to DSO? and what will be the situation for the same in the case of Infocube??
    Step by step guidance will be a great help
    Thanks in advance
    Liquid

    Hi,
    The key-fields take an important role at DSO level alone as you get the power of overwritting records. Once thats done at the DSO level you can simply carry on with a Delta load into the cube from your Change Log table and you don't have to worry about anything. Just to add, all the characteristics in the cube are key-fields so you will get a new records for each different value only the key-figures will sum up for a set of all the same characteristics.
    Thanks,
    Arminder

  • Date and Time dimensions

    After reading the following article, I have decided to use SSAS dimension wizard for generating our Date dimension, which creates a DATETIME PK.
    http://www.made2mentor.com/2011/05/date-vs-integer-datatypes-as-primary-key-for-date-dimensions/ 
    I have also created a separate Time dimension as granularity of an hour is required.
    The Time dimension is very simple and only contains a surrogate key (INTEGER) and actual time in hours (VARCHAR).
    DimTime(TimeKey, TimeInHours)
    Our Fact table will now have a link to both the Date and Time dimension using the PK's.
    Our analysis is required by hour, day, week, month and year.
    My query is; Will this current structure cause any problems when creating MDX scripts to analyse our data (i.e. drilldown and rollup queries) Hour - Day - Week - Month - Year

    Hi Darren,
    According to your description, there a day and hour granularity in your fact table, so you want to a hierarchy like Hour - Day - Week - Month - Year, right?
    In your scenario, you created a time table that only contains a surrogate key (INTEGER) and actual time in hours (VARCHAR). We cannot create a Hour - Day - Week - Month - Year hierarchy without ant relationship between date table and time table. As per my understanding,
    you need create a foreigner key in time table, and join those table in the data source view, then you can create such a hierarchy. Here are some links about create time dimension, please see:
    http://www.ssas-info.com/analysis-services-articles/59-time-dimension/1224-date-and-time-dimensions-template
    http://www.codeproject.com/Articles/25852/Creating-Time-Dimension-in-Microsoft-Analysis-Serv
    Regards,
    Charlie Liao
    TechNet Community Support

  • Creating Time dimension in BW data model. - like seen in logical data model

    Hello all,
    I have been struggling with this thing and I am looking for some help from anyone on this forum.
    We are trying to create a logical data model of our bw system. We are going live next month with Student module for universities. We have multiple Infocubes and DSO and since there is so much crossing over in between them most of the reporting is done on infosets.
    One of the thing we were thinking; is it possible to create something like a common time dimension table for every infoprovider. Basically when we are providing the reports to the end user can we give them a drop down menu which gives a time frame for reporting rather than selecting.
    Example: Like can we create something which looks in the drop down like current month data, last months data, three months ago, four months ago, five months ago, one year ago, two years ago. Can we make like these data slices in our cube and deliver it to the end user?
    We have in our cube a few date infoobjects, like receipt date, decision date, cancellation date and like wise.
    Please let me know if any one has done any similar thing, it will be very helpful.
    Thank you so much in advance.

    if you add your common time dimension to your data model, first identify for each infoprovider the time against which 'current month' and other frames should be applied and map them to your dimension.
    just a question... are you not using time dimension in cubes ? ideally this should be your time dimension llinking all.
    when you use time dimension which uses 'current month' , 'current year' , you will have to address their historisation as well. (because current month now will not be so current after 2 months).
    so in data load procedure every day these values need to change (meaning drop and reload).
    and routines to populate these values based on reporting date.
    Edited by: hemant vyas on May 6, 2009 1:56 PM

  • How to join dso and ic in an Infoset

    Hi all,
            I.ve created an infoset, under the infoset i m trying to join a dso and infocube, as of now i.ve created infocube but i don find any option to create dso beneath the infoset and i don get any option to join the both, anybody can help.
           balu

    Hi,
    while creating infoset itself, there will be option of cube and DSO, give both the technical name and continue or after cube name continue there in the infoset screen drag the DSO or wise versa and join both by minimum one common field. this infoset works like intersection.
    regards,
    SUresh

  • Duplicate time dimensions

    Hello all
    I think I have a simple requirement but I don't how it works :
    I have some dimensions and a fact tables. In my fact table, I have 2 foreign key on my time dimension.
    When I import my tables and check consistency, I have errors : "duplicate foreign key. Delete it" (don't remember full sentence)
    That is the first point that I don't understand. Why can I not have more than 1 foreign key ?
    The second point is how to do to have 2 time dimensions. In the BMM, I duplicate my time dimension. But which join do I choose ? I tested with foreign key and complex join (and 2 or 1 foreign keys in physical) and I always have the same (wrong) result.
    So can anyone explain me how to configure both physical and BMM joins to use a dimension more than 1 time (without create a new physical dimension of course)
    Thanks

    Hi,
    Create Alias tables (since they are playing different roles) for time dimension in the Physical layer and join that with your fact using physical joins.
    After Importing to physical layer, if duplicate keys exists, delete the duplicate keys.
    In this scenario two keys will be exist on the fact for the time dimension, delete those and create the joins between the aliases.
    Thanks,
    Vino

  • Dynamic members in time dimension

    <p>Hi All,</p><p> </p><p>In my outline, time dimension is sparse. and i kept certainhigher level members as dynamic calc. But, when I checked thedatabase information, it shows that all members in time dimensionas stored, even though some of them are dynamic calc. Any Idea?</p><p> </p><p>Thanks in advance,</p>

    <p>OK...I poked around and found this in our Support database:</p><p><i>In the database statistics, a sparse dynamic calcmember will be counted as a stored member, whilst a dense dynamiccalc member will not be counted as stored. This is normalbehavior.</i></p><p><i>Although Dynamic Calc members are calculated on dataretrieval, they are considered stored members because new blocksare created upon their retrieval.</i></p><p><i>An implicit or explicit shared member would not qualify as astored member because they share data and calculated values fromthe original member.</i></p><p>This is a publicly accessible database for our customers onsupport, so you can contact them for access if desired.</p>

  • Disaggregate Periodic Data to Weekly Data (Modeling Time Dimension)

    Hi,
         We are using BPC MS Version 7.0. Our users plan at period level and I need to disaggregate periodic data to weekly data by looking up weeks from Time Dimension.
    Suppose if 2009.Jan  has 4 weeks I need the following result
    Periodic Data
    2009.Jan Acct1 1000$
    Weekly Data
    2009.Jan Wk1 Acct1 250$
    2009.Jan Wk2 Acct1 250$
    2009.Jan Wk3 Acct1 250$
    2009.Jan Wk4 Acct1 250$
    Can we model Time Dimension with two hiearchies to achieves this?
    Which is best way to dissaggrate? (Stored Proc or SSIS Pacakge or Script Logic) ?
    Thanks
    Raj

    An easy way to achieve this feature is to have the time dimension be weekly and do the spread on the input schedules before sending. 
    There would be two EVDREs (one refresh only of monthly data and the other send only of weekly data).  The data range of the monthly EVDRE would be input and where the user enters data.  The data range of the weekly EVDRE would be formulas (pointing to the monthly data and doing the correct division based on number of weeks in a month).  The user is inputting monthly data, but the schedule is sending the weekly values.
    The allocation will work as well, but I would suggest to create a weekly input node under each month to capture the monthly value and then spread to the weekly members.
    JAN
    |---WK Input
    |--- WK1
    |--- WK2
    |--- WK3
    |--- WK4

  • Data not loading for new Fields InfoObjects from DSO to InfoCube

    Hi Gurus -
                   I have a DataSource that is providing data to existing DSO then to Infocube. My client asked me to added couple fields to DataSource and get the data to DSO and then DSO to InfoCube.
    Here is the Old scenarion:   DataSource -> DSO -> InfoCube.
    Here is the New Scenation:
         DataSource ( Added new fields) -> DSO (added new InfoObjects for Corresponding DataSource FIelds) -> InfoCube( Added new InfoObjects to mapped from DSO)
    I added the new fields to DataSource, added the corresponding InfoObjects to DSO and InfoCube.
    I successfully loaded data from DataSource to DSO. Data is populating for the new Fields/InfoObjects in DSO.
    But when I load data from DSO to InfoCube, I don't see any data for New Fields InfoObjects in the InfoCube.
    Data from DSO to InfoCube is loading fine for the Old InfoObjects Fields but not for the New InfoObjects I added in InfoCube.
    -Sonali

    Hi,
    Why dont u debug the load through DTP debugging and check what happens to the source field and target field once it passes through the transformation. You can easily trace back where the fields are becoming blank.
    The loads which you have mentioned earlier did it have values in Added Records/Transferred Records column for the cube.
    Regards,
    Mani

  • Issue while loading of data from DSO to InfoCube

    Hi Experts,
    Can you tell me what might root casue if data is coming into DSO from R3 its correct and fine as required but while loading it to InfoCube from DSO its showing wrong data like some of Line Items that were closed were shown open in Cube AND also KF values were not right
    Also there is no Routine code involved b/w DSO and InfoCube.
    Thanks in adv .
    NP

    Hope you didnt delete some req from DSO without deleting change log . This might cause inconsistency.
    If so , delete data from dso by right click delete data  and reload .

Maybe you are looking for

  • Calling java classes using bpelx:exec (Java Embedding) Erorr

    I try to add java embedding activity. So, I add code in this activity: <bpelx:exec name="Java_Embedding_1" language="Java" version="1.4"><![CDATA[Element ele=(Element)getVariableData("/StringForSubject"); ]]> </bpelx:exec> where ("/StringForSubject")

  • Need help with update statement with multiple joins

    I've got the following select statement that is pulling 29 records: SELECT PPA.PROJECT_ID, PPA.SEGMENT1, peia.expenditure_item_id, peia.expenditure_type, pec.expenditure_comment FROM PA.PA_PROJECTS_ALL PPA, pa.pa_expenditure_items_all peia, pa.pa_exp

  • Printing QR-Code barcodes in SAP smartforms

    Hi, Is it possible to print QR-Code barcodes in SAP smartforms using system barcode font? Can anybody please let me know the font settings for QR-Code in SE73 font maintenance? A sample QR-Code lablel can be found [here|http://images.orkut.com/orkut/

  • Why is DVI displayed at startup?

    For a few seconds at startup or when starting from sleep mode, the computer displays "DVI" in a small retangular square.  After looking at this a few hundred times and waiting all those seconds, it does not look normal and has become annoying.  Any i

  • Cannot Archive and Install--help.

    I recently upgraded from Tiger to Leopard and am having a number of problems since the upgrade. Would like to Archive and Install to get a clean System folder, but that option is now blanked out and only "Erase and Install" is available. Permissions