LO's extraction and HR extraction  7.0 flow

Dear All,
Any one suggest LO's and HR extraction is it possible in 7.0
LO's extraction and HR extraction  7.0 flow Installing the Standard Business Content any one knows please reply me ASAP.
Any standard Transformations are available In BI 7.0 please tell me the procedure and what is the Patch level.(Regarding LO's and HR).
Regards
Rama Rao G

All the transformations are not available even with the latest patch
of BI Content(patchlevel--0011,Release 703),very few are
available.
You may have to create the transformations manually as per
your requirement.

Similar Messages

  • Extraction of document flow from CRM

    Hi experts,
    I am currently struggeling with a generic extractor that will load the document flow from CRM into BW. My problem is that the load takes too long, due to the number of links in the document flow. I use a SAP FM called SREL_GET_NEXT_RELATIONS to get the document links and it works fine. The problem occurs when a campaign is linked to a transaction, since the campaign at my client can have more than 10 000 transactions and those transactions is then linked to other transactions and so on. This makes extraction of the document flow very large and slow to load.
    Is there any standard way of extracting the document flow that I am missing?
    Does anyone have any experience regarding extracting the document flow from CRM and would like to share their knowledge?
    Help are appreciated!
    /Sebastian

    there's a reason why SAP does not deliver extractors for things like this
    sometimes you need to face the end users (or project managers) with the "impact" of their requests... who would ever be able to look at a document flow that contains 50 transactions each of which linked to 1000 (or more) campaigns again each of which linked to a couple of other transactions?
    anyways, you should talk to some very experienced ABAPers if you really wish to continue with this request, rethink your approach or, what I would strongly suggest, rethink your initial request to get this into BI (if your not using web reporting, reporting on this will already crash in Excel versions 2003 and lower)

  • Sql Query need to extract the Work Flow information from Hyperion Planning

    Can Any one give me the sql query to extract the Work flow from Hyperion Planning in 11.1.2.1.
    I can extract from the Hyperion Planning but it is not in required format again I need to do lot of formating. I need the information as per the flow structure is showing linke in one perticular planning unit in all coloumn format. Hence only sql query will help me to extract this kind of information. could any one provide the sql query to extract this kind of request.
    Thanks in Advance.
    Edited by: 987185 on Feb 9, 2013 10:57 PM

    If you have a look at the data models in - http://www.oracle.com/technetwork/middleware/bi-foundation/epm-data-models-11121-354684.zip
    There is the structure of the planning application tables, have a look at the HSP_PM_* tables
    Alternatively you can look at extracting the information using LCM.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • After upgrading to Lion and iTunes 10.4, cover flow not displaying song playing. Why not?

    Since upgrading to Lion and iTunes 10.4, cover flow is not displaying the currently playing song. It will simply display the first cover in the playlist. Also, the screen saver starts during playback, unlike Snow Leopard. Any ideas on how to fix? I have already repaired disk and permissions but problem persists.

    Thank you, thank you, thank you, thank you!
    I tried restarting my mac and my iphone a bunch of times, reinstalling iTunes after deleting dozens of files, and it didn't help at all.
    I re-installed X-Code (had to use Activity Monitor to kill iTunes Helper in the middle of the install) and then my iPhone synced again!

  • Tax and pricing info are not flowing into Idoc orders02-eledp04 and 05

    Hi,
    We are carrying out SRM5.5 -SUS and ECC6.0 Integration using PI 7.
    We are configuring for the MM-SUS Scenario.
    Now,to accomodate the Tax / Pricing we have installed the add on
    component LOCINSUS and we are having doubts on how to use it:
    1) Tax and pricing information are not flowing into IDOC- orders02
    (e1edp04 and e1edp05 segments are not present in the created IDOC). The
    condition print indicator in checked on ECC System and we have also
    implemented SAP note 995986.
    2)In create ASN transaction , we are not able to see the Excise Invoice
    Number and Date field in Basic data tab as expected from SAP Library
    document although we have installed the Localization Add on component.
    Please help us to reso1ve these issues,
    Regards,
    Mani
    099890-22217.

    Mani,
    Did you resolve this issue, because we are having the same issue in ECC6 with IDoc ORDERS02.
    If you did resolve this issue could you please let us know how, much appreciated.
    Bill.

  • Drag and drop a bounded task flow in a jspx page

    Hi!!
    I am using jdeveloper 11.1.2.1 I had created a Bounded task flow which contains a jsff
    page.While i am trying to drag and drop this Bounded task flow as a region in jspx page
    am getting the error as
    Cannot create a region from a taskflow containing facelets fragements in a
    jspx page.

    see the difference between JSP XML documents and Facelets-
    http://oracleadf-java.blogspot.in/2013/05/beauty-of-adf-taskflow-difference-bw.html
    both are different view technology.
    Always Remember if
    fragment is .jsff (JspXML type)-- main page must be of JSP XML type (.jspx )
    fragment is .jsff (Facelets typee)-- main page must be of Facelets type (.JSF)
    _Ashish

  • Account Determination and G/L Accounts Assignment Flow

    Hello Guru's,
    Can anyone help me in finding Account Determination and G/L Accounts Assignment Flow with T Codes. I am new to SAP SD and I want to maintain G/L accounts for my company. Thank you.
    Regards,
    Pranava

    There are mainly links, sdn thread & message, sdn wiki & doc flowing around.
    Only thing required from your side is to put afford to search them.
    Thus, kindly put an afford to search for the topics search you want and if you find any problem in that then post your query on any of the respective forum.
    For instance, follow the following links:
    - [Account Determination|http://wiki.sdn.sap.com/wiki/pages/viewpage.action?spaceKey=ERPLO&title=Account+Determination&decorator=printable]
    - Re: Account Determination
    - [Account Determination|http://wiki.sdn.sap.com/wiki/display/ERPLO/SD+FAQ#SDFAQ-2.AccountDetermination]
    Hope you are aware of rules & regulation of SDN forums.
    -Welcome and Rules of Engagement
    -Read this before posting
    Kindly corporate.
    Regards
    JP

  • HFM Data Extract in Task Flow ??

    HI,
    I have a couple of clarification on HFM Data extract in HFM Task flow,
    1) Is Data extraction possible for the predefined accounts in HFM task flow? – in which Accounts & other dimesion memebers needs to be pre-defined so that just on execution of the task flow, HFM needs to extract the data into a text file (i.e., It should not prompt to select the Accounts (as well as other dimension members) each time we extract)
    2) For Data extract purpose, we need some of the HFM members in a different name. So, we planned to create a alias table. In data extraction, Is it possible to select the member names from an custom alias table?
    3) We have to define a specific set of Account values as negative in the extracted data file which is actually positive in HFM. Also, In the extracted data file, We need to have the Account member name same as the existing HFM Account name.
    a.     For instance,
    In HFM:
    Account Member Name in HFM: VUK
    Account Value: 1000
    In the Extracted data file:
    Account Member Name in data file: VUK
    Account Value: -1000
    Is this feasible? If yes, can you explain a bit in detail (like how we can implement this in HFM)
    Thanks,
    Siva

    HI,
    I need a help in HFM Data extraction.
    We use four custom dimensions and our requirement is to get the data extract for the custom members which we need(we need only parent members in custom dimensions).
    But, by default, HFM extracts the data for all base custom members(and not for any parent custom members) - which we dont wish to have.
    Is there any possiblty to extract the data by choosing the custom members for which we need data ??
    Our's is a classic HFM Application in 9.3.1. (in Oracle DB)
    Your response will be highly appreciated !!!
    Thanks,
    Siva

  • Export and Import of mappings/process flows etc

    Hi,
    We have a single repository with multiple projects for DEV/UAT and PROD of the same logical project. This is a nightmare for controlling releases to PROD and in fact we have a corrupt repository as a result I suspect. I plan to split the repository into 3 separate databases so that we have a design repos for DEV/UAT and PROD. Controlling code migrations between these I plan to use the metadata export and subsequent import into UAT and then PROD once tested. I have used this successfully before on a project but am worried about inherent bugs with metadata export/imports (been bitten before with Oracle Portal). So can anyone advise what pitfalls there may be with this approach, and in particular if anyone has experienced loss of metadata between export and import. We have a complex warehouse with hundreds of mappings, process flows, sqlldr flatfile loads etc. I have experienced process flow imports that seem to lose their links to the mappings they encapsulate.
    Thanks for any comments,
    Brandon

    This should do the trick for you as it looks for "PARALLEL" therefore it only removes the APPEND PARALLEL Hint and leaves other Hints as is....
    #set current location
    set path "C:/TMP"
    # Project parameters
    set root "/MY_PROJECT"
    set one_Module "MY_MODULE"
    set object "MAPPINGS"
    set path "C:/TMP
    # OMBPLUS and tcl related parameters
    set action "remove_parallel"
    set datetime [clock format [clock seconds] -format %Y%m%d_%H%M%S]
    set timestamp [clock format [clock seconds] -format %Y%m%d-%H:%M:%S]
    set ext ".log"
    set sep "_"
    set ombplus "OMBplus"
    set omblogname $path/$one_Module$sep$object$sep$datetime$sep$ombplus$ext
    set OMBLOG $omblogname
    set logname $path/$one_Module$sep$object$sep$datetime$ext
    set log_file [open $logname w]
    set word "PARALLEL"
    set i 0
    #Connect to OWB Repository
    OMBCONNECT .... your connect tring
    #Ignores errors that occur in any command that is part of a script and moves to the next command in the script.
    set OMBCONTINUE_ON_ERROR ON
    OMBCC "'$root/$one_Module'";      
    #Searching Mappings for Parallel in source View operators
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Searching for Loading/Extraction Operators set at Parallel";
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
    foreach mapName [OMBLIST MAPPINGS] {
    foreach opName [OMBRETRIEVE MAPPING '$mapName' GET TABLE OPERATORS] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT)] {
    if { [ regexp $word $prop1] == 1 } {
    incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET DIMENSION OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET CUBE OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    foreach opName [ OMBRETRIEVE MAPPING '$mapName' GET VIEW OPERATORS ] {
    foreach prop1 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (LOADING_HINT) ] {
    if {[regexp $word $prop1] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Loading Operator: $opName, Property: $prop1"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (LOADING_HINT) VALUES ('');
    OMBCOMMIT;
    foreach prop2 [OMBRETRIEVE MAPPING '$mapName' OPERATOR '$opName' GET PROPERTIES (EXTRACTION_HINT) ] {
    if {[regexp $word $prop2] == 1 } {
         incr i
    puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Mapping: $mapName, Extraction Operator: $opName, Property: $prop2"
    OMBALTER MAPPING '$mapName' MODIFY OPERATOR '$opName' SET PROPERTIES (EXTRACTION_HINT) VALUES ('');
    OMBCOMMIT;
    if { $i == 0 } {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Not found any Loading/Extraction Operators set at Parallel";
         } else {
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Module: $one_Module, Object_type: Mapping";
              puts $log_file "[clock format [clock seconds] -format "%Y%m%d %H:%M:%S"] --> Fixed $i Loading/Extraction Operators set at Parallel";
    close $log_file;
    Enjoy!
    Michel

  • The value of JMIP ( inventory BED) and SEcess( JSEI) is not flowing in MIGO

    Dear Friends,
    The Value of JEC2 is flowing in MIGO but the value of JMIP and JSEI is not flowing in MIGO excise invoice tab.
    Please suggest me the possible reasons.
    Regards,
    Balakrishna R.V

    hii
    Your requirement is for Non deductable or deductable duties ??
    If its deductible , then u shld not maintain in NVV in conditions .
    IF its non dedcutible then u have to maintain in NVV , then it won't show any break up duty calcualtion in MIRO.
    Ther in MIRO it will show u total amount it taken into account of materials i.e , inventory account .
    Hope it'll help u .
    IF not revert with detailed process how u r doing with procurement type .
    Thanks

  • Exporting simple tables in pdf to Excel or word and maintain correct text/column flow.

    Even with a black and white, clear scan at high settings, my exports to Word/Excel result are poor.  Text is recognized but not in the correct sequence for editing.  I find myself using the column select (Alt key) feature and exporting columns one at a time!  When exporting to Excel, the contents of one cell in the pdf table ends up in 20 cells in Excel among many other misalignments.  When exporting to word, it's a mess of tabs/graphics/paragraph returns...etc.  I find myself selecting the text in 25 or so graphic text boxes!
    Appears that many people have the same experience but I haven't seen a solution as advertised by Adobe.  It appears so easy on video.
    Can anyone recommend a step-by-step proven method (Win 7, Acrobat X, Office 2007, InDesign CS5.5)?
    Txs!

    I am not sure about what video, but my guess is that was a PDF produced from Excel or WORD with PDFMaker. In such a case, it would have tags and all that would retain all of the information you need. When you are working with a scanned document, you have no such tag info that provides layout info (like cell locations and such). There are some 3rd party products that are said to do a better job in such cases, but often cost as much as Acrobat and only do the conversion. If you can point out the video, then maybe we can discuss aspects of what you are doing compared to the video.

  • OWB 10gR2 map returning status = Complete and Result = Null in Process Flow

    Hi,
    I have an OWB process flow which invokes an OWB map. There are 2 outgoing transitions attached to the OWB map. 1st transition is a conditinal one (on SUCCESS) and the 2nd one is unconditional transition.
    I am monitoring the status of process flow execution in Oracle Workflow monitor. After the process flow is initiated, sometimes the OWB map within shows status = complete and result = OK in workflow monitor (in tab - status). But sometimes it shows status = Complete and Result = Null. In this situation the execution is following the unconditional transition path.
    Can anyone tell me what causes the OWB map to return a Result = NULL? Also sometimes I am getting the map returning a Result = Force (with the same status = Complete).
    The configuration that I am working with is as below ---
    OWB client version : 10.2.0.1.31
    OWB repository version : 10.2.0.1.0
    Oracle Workflow Version : 2.6.4.0.0
    Database version : 10g Enterprise edition release 10.2.0.1.0
    Regards,
    Swagata

    Hi Manohar,
    1. You need to install Oracle HTTP server in a separate Oracle home.
    2.Look for the dads.conf file in the installation. Edit the file, copy the sample code and paste it below and edit the pasted part as under ----
    <Location /pls/<SID>>
    SetHandler pls_handler
    Order deny,allow
    Allow from all
    AllowOverride None
    PlsqlDatabaseConnectString <server name>:<port>:<SID >
    PlsqlAuthenticationMode Basic
    PlsqlDefaultPage wfa_html.home
    PlsqlNLSLanguage "AMERICAN_AMERICA.WE8ISO8859P1"
    </Location>
    3. STOP and START OPMN (Oracle Process Manager & Notification Server)
    4. Use the URL in your internet browser like --- <HTTP server URL>/pls/<SID>/wfa_html.home
    5. When prompted for user id and password, give the schema user id and password where the workflow engine is installed.
    Hope this helps,
    Swagata

  • AED and SED values are not flowing into Order

    Dear Gurus,
    We are using ECC 6 and TAXINJ. We did all the CIN configurations to the best of my knowledge. JMOD is flowing normally. But AED & SED are not. UTXJ is coming as also JIN6 (VAT), E Cess and S&HECess. We dont have Cess in the client's company, so Cess is not to be considered.
    If a check list can be given, maybe I can check my customization with it, since I am an SD consultant and alot of things are involving MM too. Unfortunately we dont have any MM guy with us for the time being.
    Any answers are most welcome.
    Thanks and regards.
    Venkat - Hyd.

    Hi,
    Go to SPRO > Logistics - General > Tax on Goods Movements > India > Basic Settings > Determination of Excise Duty > Maintain Excise Defaults
    Here check whether AED and SED condition types are assigned under columns A/R AED Cond and A/R SED Cond respectively. If not then assign and check.
    Also check in FV12 for the Tax Code used whether condition record has been maintained as 100 % for AED and SED conditions. If yes then check in J1ID, for "Excise Tax Rate", maintain rates for AED and SED.

  • BED and ecess - usd values are flowing in J1IG during import procurement?

    HI,
    When I am doing import procurement with CVD & ecess on CVD, Secess on CVD conditions - all the condition values are flowing in USD values even though the condition base value is copied in INR in J1IG (excise invoice at depot,
    kindly let me know why it is not floing in INR, and what are the possible causes for this.
    regards,
    archana

    Don't you do a custom MIRO ( plan costs) on the customs vendor for the customs duties ( BCD, AED, CVD and all cesses ) before doing the material receipt?

  • How can I know the end of TCP stream and/or SOAP over HTTP flow

    Hi all!
    I want to read SOAP messages over HTTP, from sniffed TCP fragments.
    How can I decide that the TCP and/or the HTTP fragmenst finished?
    For example there is a SOAP message sliced to 5 TCP packets, and the first packet contains the HTTP header and some SOAP content too. The last packet contains the SOAP XML's ending XML-tag </soap:Envelope>.
    I don't want to watch every packet's end like "if the end string is </soap:Envelope> then it is the last packet", I just simply want to know which is the last packet of that message.
    The TCP connection won't be closed after the message arrives, and let's say the HTTP header doesn't contain a Content-Length field.

    A TCP connection is just a stream of bytes. It doesn't care what those bytes are. HTTP is built on top of TCP and specifies the ability to make a request without closing the connection (HTTP 1.1 Keep Alive). So you need to understand the HTTP protocol, understand whether it's a Keep Alive connection or not, and then do the same thing a browser would to do understand when a reply has been completed and the connection is available for the next request. Otherwise, you'll just look like you're getting loads of unrelated data as you sniff the connection. Oh, and you'll probably need to understand the HTTP chunking protocol too.

Maybe you are looking for