Dispatch Event for loaded Netconnection datas

Hi All!
I made a manager component to retrieve data from database
through amfphp.
When the datas are loaded at the end of the "onResult"
function, I dispatch an event "contactLoader"
Another parent mxml component has an instance of my manager
component with a call to a function contactLoader="getData()",
and then this function calls in turn the getContacts()
function of the instantiated manager component to retrieve the
datas. Thus I am sure that the datas are correctly loaded by my
manager before to retrieve the Array.
But I would like to refresh this array in case any change has
been made in the database. The problem is that the listener is
created automaticly by the instance of my manager component and the
"contactLoader" already exists once the data are loaded once. I
need a way to reload the data in the manager , create a new event
once the data are downloaded and then my parent component can call
one more time the getContacts function.
Is it the right way to do that? I'm not skilled enouth to
solve that problem. Thanks for your help.

"badoumba" <[email protected]> wrote in
message
news:gk57nf$oak$[email protected]..
> Hi All!
> I made a manager component to retrieve data from
database through amfphp.
> When the datas are loaded at the end of the "onResult"
function, I
> dispatch an
> event "contactLoader"
> Another parent mxml component has an instance of my
manager component with
> a
> call to a function contactLoader="getData()",
> and then this function calls in turn the getContacts()
function of the
> instantiated manager component to retrieve the datas.
Thus I am sure that
> the
> datas are correctly loaded by my manager before to
retrieve the Array.
>
> But I would like to refresh this array in case any
change has been made in
> the
> database. The problem is that the listener is created
automaticly by the
> instance of my manager component and the "contactLoader"
already exists
> once
> the data are loaded once. I need a way to reload the
data in the manager ,
> create a new event once the data are downloaded and then
my parent
> component
> can call one more time the getContacts function.
I'm not sure what the issue is...you can fire a new
contactLoader event
whenever you need to.

Similar Messages

  • What are the steps for loading master data

    Hello
    what are the steps for loading master data? i want to learn about loading all master data and the steps to choose the best way to load the data.
    if anyone has documents please send me the documents i will be really greatful
    [email protected] thanks everyone
    Evion

    Hi Heng,
    Download the data into a CSV file.
    Write a program using GUI_UPLOAD to upload the CSV file and insert records.Chk the below link for example
    http://www.sap-img.com/abap/vendor-master-upload-program.htm
    Reward Points for the useful solutions.
    Regards,
    Harini.S

  • ASCP Test for Load Legacy Data

    Through to the legacy Web Adi ASCP Data in the Test of the load.
    Plan to check more than one, but the result is not correct  that there is  cause msc_operation_components not exists.
    Questions 1: msc_operation_components Table also through the Upload Web Adi BOM_Component and Routing_operation Data should be used or whether the Pre-Process Monitor Program or other Concurrent Program please tell us what is produced?
    Questions 2: If you create a Data Collect flat file data self service if you need through the OA Template Upload for msc_operation_components Dat there is no there, what should I do??

    I think you'll find the white paper published on metalink, our main avenue for publishing documentation and providing support information to customers.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by JohnWorthington:
    Hello,
    The 11.0.2 implementation doc mentions a technical "White Paper about the Data Pump" for loading legacy data into HRMS.
    I cannot find this White Paper anywhere.
    Does it exist?
    What is the easiest way to load our legacy data into 11.0.3 HRMS??
    Thanks,
    John Worthington
    [email protected]<HR></BLOCKQUOTE>
    null

  • DTP error: Lock NOT set for: Loading master data attributes

    Hi,
    I have a custom datasource from ECC which loads into 0CUST_SALES.  I'm using DTP & transformation which has worked for loading data in the past for this infoobject.  The infopackage loads with a green status but when i try to load data, the DTP fails with an error message at "Updating attributes for InfoObject 0CUST_SALES Processing Terminated" & the job log says: "Lock NOT set for: Loading master data attributes".  I've tried reactivating everything but it didn't help. Does anyone know what this error means? We're on SP7.   Thanks in advance!

    Hello Catherine
    I have had this problem in the past (3.0B) --> the reason is that our system was too slow and could not crunch the data fast enough, therefore packets where loacking each other.
    The fix: load the data into the PSA only, and then send it in background from the PSA to the info object. By doing this, only a background process will run, therefore locks cannot happen.
    Fix#2: by a faster server (by faster, I mean more CPU power)
    Now, maybe you have another issue with NW2004s, this was only my 2 cents quick thought
    Good luck!
    Ioan

  • Best Practices for Loading Master Data via a Process Chain

    Currently, we load attributes, text, and hierarchies before loading the transactional data.  We have one meta chain.  To load the master data it is taking more than 2 hours.  Most of the master data is full loads.  We've noticed that a lot of the master data, especially text, has not changed or changed very little since we implemented 18 months ago.  Is there a precedence or best practice to follow such as do we remove these processes from the chain?  If so, how often should it be run?  We would really like to reduce the amount of the master data loading time.  Is there any documentation that I can refer to?  What are other organizations doing to reduce the amount of time to load master data?
    Thanks!
    Debby

    Hi Debby,
    I assume you're loading Master Data from a BI system? The forum here are related to SAP NetWeaver MDM, so maybe you should ask this question in a BI forum?
    Nevertheless, if your data isn't changed this much, maybe you could use a delta mechanism for extraction? This would send only the changed records and not all the unchanged all the time. But this depends on your master data and of course on your extractors.
    Cheers
    Michael

  • Steps to reducing time for loading of data

    Hi
    Could any one tell me how to reduce the time for loading of records into a particular cube or Ods. For ex: iam loading some 1 lac records.It was taking for me some 5 hrs. I want to reduce the time to 3hrs or 4hrs.What are the very first steps to be considered to make fast.
    Regards
    Ajay

    Hi Ajay,
    Check the following.
    1> Any routine you have in transfer rule and update rule should not fire database select more then one time in the same code.
    2> Load Master data before transaction data.
    3> Reduce the data pack size in infopackage.
    4> Delete old PSA because you may space issue while data loading .
    5> If you are loading in ODS then remove Bex check in ODS maintenance screen if you are not doing report on that ODS.
    hope this will help you.
    Suneel

  • Help: Algorithm for Load File Data Block Hash?

    Hi guys
    I would like to understand the algorithm of a field in INSTALL for LOAD command. It is "Load File Data Block Hash". I'm concerning about it in JCOP tool + Eclipse
    (All I want to to is designing the new INSTALL for LOAD command, and I want to imitate JCOP tool)
    Can I use DES algorithm for this field?
    Thanks in advance.

    The load file data block hash is described in GP Card Spec v2.1.1 sections 6.7.6.1, 7.7.7 and C.2 all titled Load File Data Block Hash.
    The hash is a SHA-1 hash of the load file data block.
    Cheers,
    Shane

  • Status for loading end date when GR

    Hi! We are using shipment for imports. I have set the indicator 'Cumulate 'finish loading' status from inbound deliv. sts 'gds rcpt completed' for the shipment type. But on receipting the inbound delivery, the status is still not set in the shipment created (loading end date field).
    Does anyone know what else need to be done for the date to be populated?
    Regards
    SF

    No answer

  • Lock NOT set for: Loading master data attributes error

    Hi experts,
    We were encountering this error before when trying to load master data.  When we checked the system we could not find any locks at the time, and activation or kicking off the attribute change run failed again.  We finally solved the problem running FM RSDDS_AGGR_MOD_CLOSE which sets the close flag to 'X' in table RSDDAGGRMODSTATE.  I have read that it is possible this lock error happens when two change runs happen at the same time.
    My question are:
    1. is it possible to find out what process exactly "caused" the lock? the table RSDDAGGRMODSTATE does not have a reference to any object or job. I am curious as we are trying to find ways to avoid this in the future...
    2. in our case, when we could not find any locks, is running this fm the only work around? is this a best practice?
    mark
    Message was edited by:
            Mark Siongco
    Message was edited by:
            Mark Siongco

    Hello Catherine
    I have had this problem in the past (3.0B) --> the reason is that our system was too slow and could not crunch the data fast enough, therefore packets where loacking each other.
    The fix: load the data into the PSA only, and then send it in background from the PSA to the info object. By doing this, only a background process will run, therefore locks cannot happen.
    Fix#2: by a faster server (by faster, I mean more CPU power)
    Now, maybe you have another issue with NW2004s, this was only my 2 cents quick thought
    Good luck!
    Ioan

  • Process chain for loading master data attributes

    Dear Experts,
    Can any one explain the steps needed to load master data into info object through process chain.
    Thanks in advance
    Santhosh
    Please search the forum
    Edited by: Pravender on Jun 8, 2010 3:06 PM

    in bi7 this would be:
    start
    infopackage
    dtp
    attribute change run
    M.

  • SAP R3 and BI system requirement for loading Inventory data (0IC_C03)

    Hi,
    I have installed the Business content for 0IC_C03,activated the required datasources in R3 and replicated the same in SAP BI.
    However, while filling up setup tables the load is getting failed and it is giving error "No internal table space....".
    This is related to memory issues in SAP R3.Please let me know the minimum system requirement to do the above activity,
    for example RAM,parameters to be set from basis point of view,etc.The backend is MSSQL 2003 having RAM of 20GB.There is around one crore of data.
    Please let me as  soon as possible.
    Regards
    Deepak.

    Thank you Murali.
    But approximately can you tell me what is the memory requirement for SAP R3 and SAP BI for such type of applications.
    Regards
    Deepak.

  • Error when scheduling the infopackage for loading Master data attributes

    Hi,
    Iam getting the following error message when scheduling this Master data Attributes ZIP_0PLANT_ATTR_FULL..( Flexible update of Master data info objects)..
    In Data load monitor error i got this following error message.
    Error message when processing in the Business Warehouse
    Diagnosis
    An error occurred in the SAP BW when processing the data. The error is documented in an error message.
    System response
    A caller 01, 02 or equal to or greater than 20 contains an error meesage.
    Further analysis:
    The error message(s) was (were) sent by:
    Update rules
    Thanks

    HI,
    A caller 01, 02 or equal to or greater than 20 contains an error meesage This is an Idoc error. Please check the Idocs :
    1) SM37 job log (In source system if load is from R/3 or in BW if its a datamart load) (give request name) and it should give you the details about the request. If its active make sure that the job log is getting updated at frequent intervals.
    Also see if there is any 'sysfail' for any datapacket in SM37.
    2) SM66 get the job details (server name PID etc from SM37) and see in SM66 if the job is running or not. (In source system if load is from R/3 or in BW if its a datamart load). See if its accessing/updating some tables or is not doing anything at all.
    3) RSMO see what is available in details tab. It may be in update rules.
    4) ST22 check if any short dump has occured.(In source system if load is from R/3 or in BW if its a datamart load)
    5) Check in SM58 and BD87 for pending tRFCs and IDOCS.
    Once you identify you can rectify the error.
    If all the records are in PSA you can pull it from the PSA to target. Else you may have to pull it again from source infoprovider.
    If its running and if you are able to see it active in SM66 you can wait for some time to let it finish. You can also try SM50 / SM51 to see what is happening in the system level like reading/inserting tables etc.
    If you feel its active and running you can verify by checking if the number of records has increased in the data tables.
    SM21 - System log can also be helpful.
    Thanks,....
    Shambhu

  • WebAS access via Portal: Web Dispatcher required for load balancing ABAP

    Hi Folks -
    We have EP 6.0 SP18 (Java only, WebAS 6.40, Unix/Solaris).  The portal has a CI/SCS and one DI so we have a Web Dispatcher to load balance the portal servers. This works fine (and provides port 80 access).
    This portal will provide access to HTTP services from an ABAP WebAS (6.20 with 6.40 kernel, Unix/Solaris). A landscape configuration entry has been added to the portal for this ABAP system. The ABAP system has a CI and multiple app servers, all capable of handling HTTP requests.  This will also require port 80 access.
    1. Will we need an additional Web Dispatcher to load balance HTTP requests to the 'backend' ABAP WebAS system, or will the portal be smart enough to handle the load balancing itself (perhaps based on the information in the landscape configuration)?
    2. If the portal itself handles the HTTP load balancing can you point me to documentation (so I can make sure I have proper configuration)? 
    3. Are there any changes to this with NW2004s Portal (we plan to upgrade soon)?
    Thanks in advance!  Jeff

    Jeff,
    Regarding:
    Q1. If you create a system object from the "SAP system with load balancing" template in portal and configure the object to point to your CI (msg server), the LB should be handled.
    Q2. Portal load balancing is handled by the message server.  If you point a test URL to the port of your message server, you will notice that you are issued a redirect the URL of your dialog instance.  The web dispatcher is just a proxy (with some intelligence).  When a request is made to the WD, it makes a connection to the MSG server, the list of active instances is queried, a redirect is made to that instanct.  If you use WD, that connection can be proxied behind a standard URL.   If you connect directly to the MSG Server instead, you will notice your URL change, just as it does on the service marketplace.
    WDs are good for providing services, masked (proxied) behind virtual names.  If you do not want the customer to see a physical URL of the server, use the WD.  There are lots of other solutions that can do this too though such as Apache, ISA, Juniper devices, Cisco LDs.  WDs have a very low performance threshold though, especially if you use SSL. WD is a performance bottleneck and should be benchmarked to see if it is right for your application.
    Q3. No changes this architecture in 04s.
    jwise

  • Long runtime for loading PCA data into PCA ODS

    hi all,
    hope you can help me regarding this problem:
    i have been trying to load the PCA ODS (standard) from R/3 but it is taking too long so i tried to limit the size by using key fields but no luck. i have tried to load to PSA even that is taking long for around 53440 records i am unable to figure out the problem, i have checked all the transfer rules and update rules every thing is fine.
    can anyone please help me
    thanx a lot in advance

    Hi there,
    don't know if you're loading line items, but following recommendation from SAP shortened our delta loads from 1.5 - 2 hours down to 2 minutes:
    To optimize the performance of data procurement when DataSource 0EC_PCA_3 accesses line item table GLPCA in the OLTP, you should set up an additional secondary index. This should contain the following fields:
    ·         RCLNT
    ·         CPUDT
    ·         CPUTM
    Kind regards
    /martin

  • Using VBA for loading Query data into Excel workbook

    Hi all.
    I want simply load data from BEx query into Excel Wortksheet using VBA because of report formats and footer section at the end of the results.
    Any code examples, tutorials, comments, suggestions will be regarded.
    thanx in advance,
    Gediminas

    The difficalty is that I don't know the number of rows report will return. And I need my footer only on LAST page of workbook.
    Another thing I can't imagine how to do by using standart BEx functionality is to design complex column header set (merged columns, sub-columns and etc.).

Maybe you are looking for