VM04: data transfering from specification database to table MGEF

Hello everyone,
I've customised for running VM04 in IMG.
I've assigned the Hazardous Material number 11 to the Material 1666.
I've assigned the Material 1666 to the Agent P182.
But after running VM04, the data of Hazardous Material 11 which is in table MGEF (T-code: VM03) is still not covered by the data of P182 from specification database.
Could anybody look for the reason for me, please?
Thanks!

Hello
The MGEF approach as a very old one. Form legla perspective it is not useful any more.
Documentation can be found here (e.g)
http://help.sap.com/saphelp_45b/helpdata/en/35/71880186c2223ae10000009b38f984/content.htm
I assume you have performed all necessary customizing activities? Do you receive error messages ??
Furthermore may be take a look here:
Need Warehouse Management Tables in SAP
http://help.sap.com/printdocu/core/print46c/en/data/pdf/LEWM/MMWM.pdf
http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/a0e8c107-0435-2e10-a5bc-e295ed04049b?quicklink=index&overridelayout=true
It is recommended from legal perspective to use eWM and not WM any more or at least the new WM approach as described in
http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/a0e8c107-0435-2e10-a5bc-e295ed04049b?quicklink=index&overridelayout=true
With best regards
C.B.
Edited by: Christoph Bergemann on May 7, 2011 1:42 PM

Similar Messages

  • Getting data transfered from old HD to new ones, leaving 2 slots filled . .

    Hello,
    *(Below is a thread I started in another forum, but, didn't get my final questions answered (see at bottom of this post in BOLD), and am hoping for further help and suggestions. I know there is a lot to read here, but, I think this information will help other novices, such as myself. Thanks!)*
    I have 2 new HDs (each 1T) uninstalled, and I have my old HD (160 gigs) currently residing in one of the two slots in my G5. Any ideas on the best way/process to get the data transfered from the old HD to the new ones, leaving my two slots filled with the 2 new drives AND my data from the old HD transfered and intact, with a RAID0 setup?
    OR, are there any other suggestions on a good setup with the 2 new HDs??
    I'm open to suggestions.
    I always back-up on external HDs, so, is their a need for me to worry about setting up a RAID0??
    I'm doing video processing for DVD release.
    Should I just max out the 2x1T gig HDs and not worry about a HD failure??
    All Suggestions welcomed!!
    Thanks, Jack
    *FROM THE CUTTERMAN:*
    You are limited with only 2 drive slots. In this scenario what usually works best is a small fast drive (eg Velocirapter or SSD) for OSX/Applications and a large drive for video files. It is not advisable to have the operating system running from a striped array.
    Since you already have the drives, here is what you can do.
    Install one new drive and format the partition you wish to use for OSX. You may not want to use the whole drive for OSX, so make 2 partitions. Be sure that it is GPT-formatted so it can boot.
    Use CarbonCopy Cloner to copy your current OSX partition to the new one.
    Set the startup disk as the new partition.
    Remove the old drive and replace it with the second new one.
    Reboot, and partition/format new drive.
    *Hey Cutterman,*
    Thanks for the advice. That sounds like a good set-up. I'll have to research the process of doing what you've advised, but, should work out ok:
    A. Install one new drive and format the partition you wish to use for OSX. You may not want to use the whole drive for OSX, so make 2 partitions:
    1. Partitions are done under Disc Utilities correct? Never done it before; will take a look.
    B. Be sure that it is GPT-formatted so it can boot.
    2. GPT is also done under Disc Utilities, correct? Here I start to get a bit confused, as I've formatted Extended Journal, SO, the partition with the OSX & Applications is formatted GPT, and the rest of the HD is formatted as Extended Journal?
    C. Use CarbonCopy Cloner to copy your current OSX partition to the new one.
    3. So, I partition the first new HD and use Carbon Copy Cloner (also under Disc Utilities) to copy the OSX & Applications from my original HD to the first new HD, correct?
    D. Set the startup disk as the new partition.
    4. This will be an option when I setup the partition???
    E. Remove the old drive and replace it with the second new one.
    F. Reboot, and partition/format new drive.
    5. QUESTION: So, you recommend not striping the HDs: If I do a RAID 1, which sets up a mirror of the HDs, and since the first new HD is partitioned with the OSX & Applications, AND is set up as the Startup Disk, then the second new HD will also mirror that?? Another QUESTION: Even if the second new HD does not mirror the OSX & Applications, the second HD will only mirror the space/partition not utilized by the OSX & Applications partition, correct?
    Thanks for your help!!
    Jack
    *FROM THE CUTTERMAN:*
    OK I will try to answer your questions. To begin, if you want to mirror the 2 drives then you will need to boot from the SL install DVD or an external (USB/firewire) drive to set it up. IMO for your purposes it is too much hassle and a waste of HD space.
    1) Yes partitions are created in Disk Utility
    2) Yes, you choose a volume Scheme (ie number of partitions) and size them by dragging the separator bar. Then select the planned boot partition and click Options..., then choose GPT. I think it is usually the default but check and make sure. Extended journaled is the usual format.
    3) You need to download Carbon Copy. It is a free tool that will copy the entire image of your current system partition to the new one and make it bootable. Consider making a donation as it is a very useful and frequently updated utility. It is fairly intuitive to use. You can also use the restore feature in Disk Utility but I have had more experience with Carbon Copy.
    4) Once the copy process is finished the new boot partition will show up under the startup disk selections (this tool is in System Preferences).
    *FROM 666Sheep:*
    If may i correct one thing: OP, don't choose GPT (GUID Partition Table). You got PPC Mac (G5), so valid partition type for you is Apple Partition Map (APM).
    GPT is for Intel Macs and you will not able to boot from this kind of partition.
    *FROM THE CUTTERMAN:*
    Thanks for the correction- my bad. No familiarity with non-Intel Macs.
    G5 PowerMac, 2.5 GHz Dual, Dec. 2004, 6.5 GB RAM, 149 GB HD
    *Hi All,*
    *Since this process will come to a head on Tuesday/Wednesday, and I hope to be successful in transferring Old HD data to new HD/s.*
    *A: I'm still a little foggy here, as I thought I'd set up the 2 new T-gig HDs so that if one of them failed, I'd still have the data backed up on the other HD; more ideas concerning this would be reassuring:*
    *WHAT DOES "IMO" mean?*
    *(IMO for your purposes it is too much hassle and a waste of HD space.)*
    *As I understand the points made: After I partition the first new HD, and Carbon Copy data from old HD, I then install second new HD, and this will just act as an overflow from the first partitioned HD, correct? Because, if set up a Raid1/Mirror, the OSX partitioned portion of the first new HD will not copy to the second new HD, and that space on the second new HD will be wasted, correct?*
    *Also: I have to set up RAID1 (if I do not set up RAID0), correct?*
    *RAID 1*
    *From WIKI: RAID 1 mirrors the contents of the disks, making a form of 1:1 ratio real time mirroring. The contents of each disk in the array are identical to that of every other disk in the array. A RAID 1 array requires a minimum of two drives.*
    *Carbon Copy is a free tool that will copy the entire image of your current system partition to the new one and make it bootable.*
    *1. Is my old HD partitioned? Is that done automatically by MAC prior to PC purchase?*
    *2. Do I Carbon Copy the whole old HD or just parts? (eg. OSX, Photo Shop, After Effects, Various Files)*
    *3. I read that it is necessary to DEACTIVATE Photo Shop (I have CS3) prior to doing Carbon Copy. Is that correct? If so, does this also apply to other Adobe applications? Such as After Affects, Illustrator??*
    *Thanks to all for the help!!*
    *G5 PowerMac, 2.5 GHz Dual, Dec. 2004, 6.5 GB RAM, 149 GB HD*

    Confusing Information:
    A1: You also want to make sure the drive you are backing up to is formatted Mac HFS Extended (HFS+) if using Mac OS 8.1 or above.
    A2: On PowerPC Macs, your clone should be partitioned as Apple Partition Map.
    *Q? I thought the new HD (1T) should be formatted in Extended Journaled?*
    B: Also disable Spotlight (in 10.4 only) on your destination drive using Apple menu -> System Preferences -> Spotlight -> Privacy to add the destination drive to the pane.
    *Q? I'm using 10.4.11, so, I should follow these instructions?*
    C: If possible, boot into safe mode to perform the backup (holding the SHIFT key at startup). In addition, you can clone while logged into another administrative user that you don't use at all to avoid further complications of changes which may be happening to your regular user (though don't use Fast User Switching to get into that other user, since that other user is still active when fast user switching is used). Otherwise you are going to be running a backup on a live system which could have changes happening while you are attempting to backup. These may yield an imperfect clone, with uncertain success at recovery. It may be possible that your clone will have its own hardware issues, so make at least two copies.
    *Q? ?????????????*
    *I GOT THE ABOVE INFO. FROM THE FOLLOWING SITE:*
    Making a clone/mirror/duplicate backup
    http://www.macmaps.com/backup.html#SHORTANDEASY
    *ALSO, WHAT DO YOU THINK ABOUT THE ADVICE ON THIS PAGE:*
    http://www.levoltz.com/2010/04/21/how-to-transfer-data-to-your-new-hard-drive/

  • Data transfer from oracle database(external database) to sap database

    Hi Abapers,
    I have a requirment like excel data upload from oracle database(external database) to SAP database .Plz any one can help me how to create a bdc program for this requarment.
    Advance thanks you.
    Jnana

    HI
    use function module 'ASLM_EXCEL_TO_INTERNAL_TABLE' TO UPLOAD DATA FROM EXCEL SHEET TO INTERNAL TABLE. FROM THERE U CAN SAVE IT TO DATABASE USING BDC FUNCTION MODULES.

  • Doc type and drilldown Consolidate data transfered from FI

    Dear all members,
    I config doc type for data realtime update from FI.
    But when i run report CX34, doc type of data transfered from FI to EC-CS is blank.
    And I can't drilldown this data.
    Please show my mistakes
    Many thanks,
    Orchid
    Edited by: Orchid675 on Jul 2, 2010 5:12 AM

    Dear Orchid,
    in EC-CS the document number and the document type will be new, the doc type info will not be updated so the original FI doc type will disappear.
    Typically, in EC-CS all reported financial data (so documents from FI) will be updated with the same doc type at posting level SPACE (realtime update) or 00 (all other).
    I hope my question helps.
    Ciao,
    David

  • How to delete Master Data transferred from ECC to APO via CIF

    Hi, all
    I'm connecting SCM 5.1 with ECC 6.0 via CIF.
    When activating Integration Model in ECC, the following error occurred;
       "Location 0001 of type 1001 (BSG 592BUSGP) already exists in APO"
    This is because Location 0001 was already transferred from different ECC client via CIF, and not cleared.
    I'm now trying to initialise master data transferred from ECC to APO via CIF, but in vain.
    I suppose I have to first delete all master data CIFed from ECC in APO, or have to execute initialisation of APO master data with ECC data in ECC.
    Please tell me how to do this.
    For details, please read the followings;
    I connected APO client with ECC client 590.
    I transferred Plant and Material Master from ECC to APO via CIF.
    After that, I wanted to change ECC client from 590 to the other.
    I created the other client (592) by client-copying from 590 with SAP_ALL copy profile.
    I deactivated Integration Model in ECC 590, and activated the same in ECC 592.
    Here, I faced the problem.
    The initial data transfer failed due to the duplicate of data.
    Now I'm wondering how I can initialise master data in APO and transfer initial data from ECC.
    Due to testing purposes, I have to change ECC clients from one to the other this time, but I cannot initialise master data in APO and cannot tansfer initial data from different ECC client.
    thanks in advance,
    Hozy

    Hi all,
    Thank you very much indeed for lots of your replies.
    Thanks to all of your advices, I'm getting to understand what to do.
    Well, I'll say I have to create one more client for testing with other ECC clients or I have to rebuild the client by client-copying from 001 and implementing all transportation requests, for marking del flags to each master data and running del programme sound like tough work, and sound like odd ..
    Then, I've got next issue here.
    In SCM-APO, I use BI objects, and , if I have to create several clients for testing, I have to set up clients where you can use both BI objects and CIFed master data.
    I suppose I have to change 'BWMANDT' in RSADMINA and make RSA1 accessable in other SCM clients and implement tests.
    Also, I have to register BI System for newly created SCM clients in RSA1-SourceSystem.
    Well, if you have any experiences on this kind of use, and if you have any advice on it, please give me any tips.
    Particulary, I'd appreciate very much if you give me advice on which way you have chosen; creating several clients or rebuilding the one client in SCM-APO when you use both BI objects, CIFed master data and Demand Planning.
    Thanks in advance.
    Hozy

  • Can I use stored procedures to improve data retrieval from the database?

    I have been requested to get multiple pieces of data from multiple tables and return the output in an array for use in a Java program.
    Eliminating multiple calls to the database. Currently, the data is retrieved from the database in multiple steps but to eliminate
    possible bottlenecks in the future, I created several complex sql statements.
    For instance,
    In the first table there are 20 columns of data. has a column named type which can have the one of these values: line, inn, or bsc
    In the second table there are 6 columns of data. has the additional data for type line
    In the third table there are 7 columns of data. has the additional data for type inn
    In the fourth table there are 2 columns of data. has data regardless of type
    etc...
    Depending on a specific column value retrieved (inn) from the first table:
    I need to collect all 20 columns in the first table and get all the columns in third table and the columns in the fourth table, etc.
    Question: How can I remove the duplicate primary key columns without specifying for each table the column names to be returned?
    I used the wildcard SELECT * because I didn't want to have to update the retrieval methods if additional columns are added to these
    secondary tables.
    example of my sql for type inn data:
    select * from first_table f, third_table s, fourth_table t
    where (f.column1='xxxx' and f.column2='xxxx') (the 'xxxx' are passed into the sql statement)
    and
    ((s.column1=f.column1 and s.column2=f.column2)
    and
    (f.column3=t.column1));
    Currently, I've duplicated the separate sqls for each valid type found in the first table (at least a dozen types). The only difference is
    the table name. Instead of the third table I use the second table, etc.
    The MAIN Question: How can I set this up to have one sql to handle each type? I want to eliminate having over a dozen duplicated sqls. Can I
    incorporate this into a stored procedure or something? If so, how? Can anyone provide sample coding?
    Your help is very much appreciated. Thank you.
    GD

    Hi, Salah.
    Use "Exec" in your query to run procedures.
    SAPbobsCOM.Recordset     oRS;
    oRS = (SAPbobsCOM.Recordset)pCmp.GetBusinessObject(SAPbobsCOM.BoObjectTypes.BoRecordset);
    oRS.DoQuery ("EXEC YourStoredProcName");
    Triggers are not supported in SDK.
    Regards,
    Aleksey

  • Data extraction from Oracle database

    Hello all,
    I have to extract data from legacy database tables. I need to apply a lot of conditions on data extraction using SQL statements for getting only valid master data, transaction data, SAP date format etc. Unfortunately I don;t have a luxary of accessing legacy system data base table to create table views and applying select statements.
    Is there anyother way round by which I can filter data on source system side w/o getting into legacy system. I mean being in BW data source side.
    I am suppose to use both UD connect and DB connect to test which will workout better way of data extraction. But my question above should be same in either interface.
    This is very urgent as we are in design phase.
    Points will be rewarded immediately.
    Thanks
    Message was edited by:
            Shail
    Message was edited by:
            Shail

    Well I and eveyone know that it can be done in BI.
    I apologize that I did not mention it in my question.
    I am looking for very specific answer, if there is any trick we can do on source system side from BI. Or where we can insert SQL statements in infopackage or data source.
    Thanks

  • Data Extraction from SAP R/3 Table View in BW

    Dear All,
    I am trying to create a BW query from SAP R/3 Table View.
    Till now i have created a table view
    Extracted Data source in R/3 Useing Tcode : RSA3
    In BW Side
    1. I have replicated the data source in Source system
    2. Created  info object
        and info object catalog
        and assigned it
    Now i am little confused over what to do next
    wheather to create ODS object in info cube,
    Please guide me with rest of steps to follow.
    Regards,
    Gaurav

    Hi,
    After replicating DS in BW, you have to maintain ODS or Cube for maintaining data in BW.
    After loading data into Data Targets you can create Reports based on those Data Targets.
    For better reporting performance you can create Cube rather ODS.
    Info Source(After Creating info Objects, you have to create a Transfer Structure to map the Fields.)
    Then create a Data Target ODS/Cube.
    Update rules
    Info Package
    Hope it helps you .
    Let us know if you still have any issues.
    Reg
    Pra

  • Data transfers from SRM to R/3

    Hi,
    i have follwing questions?
    1. How date transfer from SRM to R/3?
    2. Why PO transfer throught RFC? why not IDoc? and why confirmation and invoices are transfer using IDOC?.
    3. config document for EBP-SUS?
    Thanks
    Rams

    Hi Rams,
    The data transfers between SRM using ALE (Application Link Enabling) via RFC & IDOCs...
    P.O is transferred to R/3 using RFC. Confirmation and invoice through IDOCs.
    Since P.O is the leading document and of course legal document, there should not be any duplication. Means if it generates two P.O numbers one in SRM & the other in R/3 they are considered as two different P.Os.And will be an issue during audit.
    Where as confirmation and invoice posting documents are internal documents and the docs created in the back end with the back end number range. In this case same number is not mandatory.
    Hope this makes you more clear.
    Award points for helpful answers.
    Rgds,
    Teja

  • IReport data retrieval from MySql database

    Hi,
    can anyone help me? i am using iReport - jasperreports to design a report using a datasource/ connections.
    i designed my report and everything works fine except that the data of the fields in mthe detail band is been retrieved more than once making the reports to have one data printed more than once(as many as possible). In my quey, i even used the "order by" as specified in the iReport manual.
    Questions: Is there to fix this so that i can have one data of a field printed only once witout repeatition in the detail band(that is, print one value then the next till the last value then stop)

    Play with the evaluated and printed when options. Another thing you should do is create a group whose evaluation string is the key from your database. every time this value changes as the records are being evaluated the detail section will be printed.

  • Data needed from emp and dept tables

    Wondering if somebody can querry the emp table and dept table that comes with some versions of oracle already built in.
    I need the data produced from these two querries
    select * from emp
    select * from dept

    If you look in ORACLE_HOME/sqlplus/demo you'll find demobld.sql which contains the script the build all the scott tables.

  • Data access from Access Database

    I want collect some data from access, I have created a frame with swing componet, now when I am searching data with some variable from Access database , it is not comming, but in simple case all data are comming in differnt places in my frame.
    The string which I used is
    st.executeQuery("Select * from student where cid = '+st1+'"
    where st1 is avariable collected some data from the frame, it is tested ok.
    now i never did practice with access, please if there is any wrong .

    Could you include more code please? The single statement you have provided seems fine, but we're not psychic.

  • SEM-BCS data extractor from ECC general ledger table(s)

    We are a utility company working on an SEM-BCS implementation and use the FERC solution.  We do not use the new GL.  We are trying to extract the transaction data from ECC to a BI virtual remote cube.  We cannot use the profit center extractor (0EC_PCA_3) as the profit center tables do not contain any ferc data.  We need to be able to extract the transaction data from a general ledger table.  We have run into several issues with various extractors we have tried because they donu2019t allow direct access (0FI_GL_4) or are at a summary level and we canu2019t extract group account, trading partner, and transaction type detail (0FI_GL_1).  Would you have any suggestions on how to extract general ledger data with the detail information required from ECC to be able to load to a BI virtual remote cube?

    We are going forward with getting the natural account detail data using the profit center extractor 0EC_PCA_3, and getting the ferc summary data using the general ledger extractor 0FI_GL_1.  With our testing so far, this combination will provide us the data we need in BCS.

  • Data Transfer from Oracle database 10g to Oracle database 9i

    Hi Experts,
    We need to insert records at the speed of min 10,000 records per sec in the following condition.
    Source : Oracle 10 g and Oracle 9i
    We need to select the data from Oracle 10g and insert into Oracle 9i database.
    Here we not allowing to create database link and also not allowing to create view or materialized view
    because this two database is not on same network.
    So we developing the small java application on intermediate server where we write process to get the connection of this two database servers.
    From java application we call the procedure for selecting data from Oracle 10g and insert into oracle 9i database.
    What is the best way to achieve?
    Or You also suggest any other way to achieve
    ( As per this scenario materialized view is working ? )

    Thanks freiser,
    But it create another problem as per my business logic, There will be two database server , one is online server where online user fill the form which is generated by java, spring , hibernate and using database 10 g.
    at day end i need to execute a process that transferring data from online server to offline server that is in oracle database 9i. This process is scheduled. Some security reason client do not kept this two database on same network.
    My challenge is that transfer data from online server to offline server with applying client security norms.
    I have option like
    1) Using Oracle replication method, creating materialized view on remote server , refreshing it at regular interval. but database connectivity is not contineous, should i go for that ?
    2) Write java application on intermediate server where we write process to get the connection of this two database servers. From java application we call the procedure for selecting data from Oracle 10g and insert into oracle 9i database and using flag on both data to identified how many rows are transfered and how many remaining for trasfer.
    Please tell me what is best way to acheive this ?

  • HELP! Displaying results from a database in table format

    I'm developing a web application that will use a considerable number of database queries that need to be displayed back to the web-browser in HTML table format.
    I have been told that that result sets can allow tables to be built automatically without having to write the same loop and display code over and over.
    So far I have a HtmlResultSet class as follows:
    import java.sql.*;
    public class HtmlResultSet {
    private ResultSet rs;
    public HtmlResultSet(ResultSet rs) {
    this.rs = rs;
    public String toString() {  // can be called at most once
    StringBuffer out = new StringBuffer();
    // Start a table to display the result set
    out.append("<TABLE>\n");
    try {
    ResultSetMetaData rsmd = rs.getMetaData();
    int numcols = rsmd.getColumnCount();
    // Title the table with the result set's column labels
    out.append("<TR>");
    for (int i = 1; i <= numcols; i++) {
    out.append("<TH>" + rsmd.getColumnLabel(i));
    out.append("</TR>\n");
    while(rs.next()) {
    out.append("<TR>"); // start a new row
    for (int i = 1; i <= numcols; i++) {
    out.append("<TD>"); // start a new data element
    Object obj = rs.getObject(i);
    if (obj != null)
    out.append(obj.toString());
    else
    out.append(" ");
    out.append("</TR>\n");
    // End the table
    out.append("</TABLE>\n");
    catch (SQLException e) {
    out.append("</TABLE><H1>ERROR:</H1> " + e.getMessage() + "\n");
    return out.toString();
    I also have created a class that includes an instance of the class above to display the results, as follows:
    import java.io.*;
    import java.sql.*;
    import javax.servlet.*;
    import javax.servlet.http.*;
    public class customerLookup2 {
    public void doGet(HttpServletRequest req, HttpServletResponse res)
    throws ServletException, IOException {
    Connection con = null;
    Statement stmt = null;
    ResultSet rs = null;
    res.setContentType("text/html");
    PrintWriter out = res.getWriter();
    try{
    Class.forName("org.gjt.mm.mysql.Driver");
    con = DriverManager.getConnection("jdbc:mysql://localhost:3306/sandywalker");
    out.println("<html>");
    out.println("<head><title> Users</title></head>");
    out.println("<body> Customers");
    HtmlResultSet result = new HtmlResultSet ("SELECT * FROM USERS");
    out.println("</body></html>");
    catch(ClassNotFoundException e ) {
    out.println("Couldn't load database driver: " + e.getMessage());
    catch(SQLException e) {
    out.println("SQLExeption caught: " +e.getMessage());
    finally {
    try {
    if (con !=null) con.close();
    catch (SQLException ignored) { }
    I keep getting a compile error "customerLookup2.java": Error #: 300 : constructor HtmlResultSet(java.lang.String, java.sql.Connection) not found in class HtmlResultSet at line 42, column 34".
    Can anyone shed any light on this.

    Looks like you are passing a string (SELECT * FROM USERS) instead of passing a ResultSet. Try passing a Resultset.

Maybe you are looking for

  • Windows 7 error code 0x80070570 during installation.

    Hi there. I currently tried to reinstall my OS, so I did. I finish the expanding files part, but then I get the error message. I've looked at other posts, but I can't understand a word you guys say. Im not smart in computer so yea. I see a soultion t

  • ATV limited to 720p, 24 fps?

    There are many H.264 camcorders nowadays like the Flip MinoHD, VadoHD, Sanyo Xacti and the Kodak Zi6 among others. By default, recording is usually at 720p at 30 fps. The ATV cannot play back these recordings without having to downgrade the format to

  • Manual kerning/tracking in Incopy looks different in Indesign

    Does Incopy handle kerning and tracking differently than Indesign? Editors are opening INDD file in Incopy and using manual kerning and tracking to copyfit. However, when the INDD file is opened in Indesign, the kerning and tracking appear not to hav

  • DHCP scope full, event ID 1020

    Hi, one of our Windows 2008 R2 Domain controllers is returning the following warning message on almost a daily basis: Log Name:      System Source:        Microsoft-Windows-DHCP-Server Date:          19/11/2014 11:32:41 AM Event ID:      1020 Task Ca

  • WebAPI question about Hierarchy modeling

    Hi all, I have come to a issue about Hierarchy modeling. I want to use WebAPI to drill down the query in the visual composer. And I have follow every steps mentioned in this PDF file </people/prakash.darji/blog/2006/08/24/overcome-limitations-around-