Issues with removing Flex data binding to use Air 14

Hello,
I was wondering if anyone had optimal solution for replacing the Flex data binding classes in Air 14. I'm unable to merge the Flex and Air SDK together anymore so all the binding handlers need to be replaced. Has anyone else run into this issue yet? The automated binding generation and handling was the best feature of Flex and now that it's broken it creates huge issues for me.
This is for Air desktop and mobile applications. They still build just no of the event handling works.
Cheers,
Pete

I had forgotten I solved this myself last Dec. So basically from what I can tell the precompiler no longer does any [Bindable] conversions so you have to manually create the getters and setters for your model when this changes. I hope this helps anyone who may run into the same problem. There's literally no documentation on this any wheres.
i.e.
[Bindable]
public var username : String;
becomes
private var _username : String;
        [Bindable(event="propertyChange")]
        public function get username():String
            return this._serviceState;
        public function set username(value:String):void
            var oldValue:Object = this._username;
            if (oldValue !== value)
                this._username = value;
                if (this.hasEventListener("propertyChange"))
                    this.dispatchEvent(mx.events.PropertyChangeEvent.createUpdateEvent(this, "username", oldValue, value));
You also need to make sure what ever class contains this property implements IEventDispatcher and has the following functions:
IEventDispatcher implementation
private var _bindingEventDispatcher:flash.events.EventDispatcher =
new flash.events.EventDispatcher(flash.events.IEventDispatcher(this));
  * @inheritDoc
public function addEventListener(type:String, listener:Function,
  useCapture:Boolean = false,
  priority:int = 0,
  weakRef:Boolean = false):void
_bindingEventDispatcher.addEventListener(type, listener, useCapture,
priority, weakRef);
  * @inheritDoc
public function dispatchEvent(event:flash.events.Event):Boolean
return _bindingEventDispatcher.dispatchEvent(event);
  * @inheritDoc
public function hasEventListener(type:String):Boolean
return _bindingEventDispatcher.hasEventListener(type);
  * @inheritDoc
public function removeEventListener(type:String,
listener:Function,
useCapture:Boolean = false):void
_bindingEventDispatcher.removeEventListener(type, listener, useCapture);
  * @inheritDoc
public function willTrigger(type:String):Boolean
return _bindingEventDispatcher.willTrigger(type);

Similar Messages

  • Issue with Adobe flex data.xml file not reachable from bsp behind firewall

    Hi Gurus,
    I have a problem with the <mx:HTTPService> tag the following is the actual tag,
    <mx:HTTPService
    id="Srv"
    url="data.xml"
    useProxy="false"
    method="POST" result="resultHandler(event)"/>
    When accessed locally I can see the data in the flex as the data.xml can be reached. when the same is accessed from internet behind firewall, the url is entirely different and the .swf file in the BSP page cannot access the data.xml. I cannot give the absolute url in the tag as the BSP page application is accessed differently in different servers. any help on this would be greatly appreciated
    Thanks
    Akbar

    Sorry somehow I missed this question, an even easier way to do this is to allow your Flash movie to "script" ( this is the default behavior for a Flex application ) and then call some javascript to obtain exactly what the page URL is and then go from there to get your data:
    import flash.external.ExternalInterface;
    import mx.utils.URLUtil;
    var
    if(ExternalInterface.available){
         pageURL = ExternalInterface.call("window.location.href.toString");
    // Do whatever you need with the URL here.
         var serverName:String = URLUtil.getServerNameWithPort(pageURL);
    -d

  • Issue with Ord Start Date

    Hi,
    I am having issue with Ord Start Date.
    When i create XYZ Asset now when i do the acquasation for that new asset. Example: Using FB01 i enter todays date and i do the posting.
    Now when i go to DEP Area TAB for the DEP Key (M200) - The Ord Dep Start Date is 15 Dec 2010.
    For this DEP Key Period Control Methods (T code - AFAMP) - Acquisation Entered is ( 6 that is - at the start of year) and the Period Control is ( T code - OAVS) (4 that is - First year convention at half year start date).
    Our client is using the fiscal year as 1 June 2010 to 31 May 2011.
    So when i do acquistion for the asset as per the period contol the Ord Dep Start Date should be 1 Dec 2010, but here it is talking 15 Dec 2010.
    Why is system talking 15 days more.
    Regards.

    Hi
    Which Fiscal year variant are you using??
    Did you maintain it in OAVH? Copy the existing entry in OAVH for K4 to your Fisc Yr var and try again
    And also, 04 does not figure for Acquisitions in AFAMP under 0004.... Whatever you specify in OAVH must also be a part of AFAMP
    Regards
    Ajay M
    Edited by: Ajay Maheshwari on Oct 28, 2010 2:31 PM

  • Issue with list saving data after sites upgrade from sharepoint 2010 to sharepoint 2013

    Issue with list saving data after sites upgrade from sharepoint 2010 to sharepoint 2013 
    Newform.aspx of list:-
    Custom List is not saving data sometimes in the new form after 15 minutes and only blank entry record got created without saving data, even though some columns are mandatory fields?

    Hello dcakumar,
    Sounds like a strang issue. If you can reproduce this can you see some errors in the ULS logs?
    - Dennis | Netherlands | Blog |
    Twitter

  • Issue with Removing the Patch Set Software 9.2.0.8

    Hi experts,
    I have issue with removing the patch set software 9.2.0.8
    I'm working on solaris box recently ugraded to 9.2.0.8 from 9.2.0.7. however i want to downgrade now. I have followed the document 4547809 mentioned. Now the real issue is when i restored the oracle_home (9.2.0.7) and execute sqlplus /nolog here is error message shown
    bash-2.05$ sqlplus /nolog
    bash: /oracle/test/920_64/bin/sqlplus: No such file or directory
    bash-2.05$ lsnrctl
    bash: /oracle/test/920_64/bin/lsnrctl: No such file or directory
    Please help
    thanks in advance

    Hi,
    Thanks for reply you are right
    ls -l /oracle/test/920_64/bin/sqlplus was not exist. However i copied everything from the backup.
    I'm trying again
    cp -r bk_test_home/* /oracle/test/920_64/*
    still i couldn't the file however it exist in bk_test_home. please suggest
    Thanks and regards

  • My debit card kept getting rejected by my iTunes account, it there an issue with iTunes? I have been using the same Card since very long time. I downloaded the iOS 8, can I talk to someone so I can get this resolve?

    My debit card kept getting rejected by my iTunes account, it there an issue with iTunes? I have been using the same Card since very long time. I downloaded the iOS 8, can I talk to someone so I can get this resolve?

    Apple is not here. Contact Apple to resolve your issue:
    iTunes Support WorldWide
    or
    iTunes Support by eMail ww
    Cheers,
    GB

  • Performance issue during SharePoint list data bind to html table using Ajax call(Rest API)

    Hello,
    I am having multiple lists in my SharePoint Site. I am using SharePoint REST APIs to get data from these lists and bind a HTML Table. Suppose, I have 5 lists with 1000 records each, I am looping 5000 times to bind each row(record) to this html table. This
    is causing performance issue which is taking a very long time to bind. 
    Is there any way So that I can reduce this looping OR is there any better approach to improve the performance. Please kindly Suggest.  Thank you for your help :)
    Warm Regards,
    Ratan Kumar Racha

    Hi Racha,
    For handling large data binding in a page,
    AngularJS would be a great option if you might would worry about the performance.
    You can get more information about using AngularJS from the two links below:
    https://www.airpair.com/angularjs/posts/angularjs-performance-large-applications
    http://www.sitepoint.com/10-reasons-use-angularjs/
    Best regards
    Patrick Liang
    TechNet Community Support

  • Issue with 0hrposition master data

    We are extracting data from SAP using 0HRPOSITION_ATTR datasource. I noticed that data is not maintained correctly in master data tables in BW and giving us incorrect results in reporting. Consider below mentioned scenario:
    Position A created as vacant on 04/01/2006 with start date (BEGDA/ Valid from) as 04/01/2006 and End date (ENDDA/ Valid to) as 12/31/9999 and is vacant. Below mentioned entry is shown under maintain master data for 0HRPOSITION in BW:
    Position Valid To Valid From Position Vacant
    A 03/31/2006 01/01/1000
    A 12/31/9999 04/01/2006 X
    Position A is now delimited on 09/15/2006 as it’s no more required. In SAP, position has record only from 04/01/2006 till 09/15/2006 as vacant. When record is extracted in BW, it creates below mentioned entry in master data table.
    Position Valid To Valid From Position Vacant
    A 03/31/2006 01/01/1000
    A 09/15/2006 04/01/2006 X
    <b>A 12/31/9999 09/16/2006 X</b>
    Entry 09/16- 12/31 is incorrect as position doesn’t exist for this duration. If we report on 0HRPOSTION with key date as 09/30/2006, it shows position A as vacant though position no longer exists.
    Has anyone come across this situation. any help is greatly appreciated.
    Kamal
    P.S: Milind Rane...I was searching through the forums and came across your post.I would appreciate if you could let me know how you solved this issue...
    Message was edited by:
            Kamal K

    HI KK
    I have a similar issue. Can you please let me know how this Issue with 0HRPOSITION_ATTR extractor was resolved. In my case i have incorrect data in BW when reporting is done. THrough the extractor untill the PSA i have the correct records coming  but in the master data 0HRPOSITION i have incorrect records.
    Please help.
    Thanks
    Hari
    Message was edited by:
            SAPCOOL

  • How can I create a sample with the Flex data?

    Hi mates, I'm thinking about doing some stutter vocals with the ultrabeat so I've got a vocal and I have flexed it in time so it fits well on the song's tempo, the problem is that when I create an audio file of the part of the vocal I want to use on my ultrabeat, it creates an audio file that's not fitted in time.
    How can I create the audio file fitted in time?with the flex info?
    Thanks so much!

    psikonetik wrote:
    How can I create the audio file fitted in time?with the flex info?
    You need to create Apple Loop, REX etc.
    For example select the region that you have already flexed, right click ->Bounce Merge->Bounce Inplace. I will create a new audio track with the bounced Flex work.
    Select the new region and go to Audio menu ->Open in Apple Loop Utility and check "Loop" and other attributes to create an Apple Loop. Save the Apple Loop and close the utility. It's expected that the Apple Loop aif must be created in your Project path folder or have a look at the Audio Bin in Logic where is the aif path.
    Drug the Apple Loop in the Arrange audio track ( or it will create one ) and operate with any tempos.
    !http://img59.imageshack.us/img59/4967/aglogo45.gif!

  • Issue with extraction of data to Delta Queue

    Hi All,
    Background:- One LO Data source each from Application 8, 11 and 12 was active in LBWE. Update method was specified as Queued Delta. However this Data source was not utilized to pull any data to any of the Global BI Systems. These data sources did not exist in RSA7. There was a business requirement that these Data sources were to be enhanced, along with this other Data sources under these applications were also required to be enhanced.  There were entries in LBWQ which used to get dumped out at regular interval by running the SAP Standard Collector jobs. Since no Delta was initialized for these extractors, no entries were passed on to RSA7.
    Problem:- We did not clear the entries or took a downtime before sending transports and now the collector job is dumping saying the structures have changed, which is correct.  I believe we wonu2019t require any data sitting in LBWQ as none of the extractors had any entry in RSA7. This problem is happening only in our Production system and we did not face this problem in any other system even though there were entries existing in LBWQ.
    What is the solution to fix this? Is the only option now is to delete the Queue from LBWQ and then run the collector job or is there any other way and will that fix the problem?

    Hi Victor,
    As u said that u were not using the DS for Application 8,11 and 12.. in this case I believe that u dont have any data in BW as well(Historical data). As u have enhanced these data sources with the new requirement, I hope u must have changed the structures in BW side as well to load the data to the newly added fields.
    So in ur case u need to replicate the DS in BW and then activate the Transfer structure for all of them.
    And then flushout the LBWQ for all those 3 applications and then take the setup tables fill for the same (but this requires the down time in prod to make sure that u wont loose any records).
    Before setup check SM13 whether u have any blocked entries for all these entries if so then u can delete them as well before taking the setup.
    Once the setup is done then u can take Init without data transfer and then take repair full loads in BW for the targets.(Activate ur V3 once the init- without data transfer took place for the regular deltas) .
    Thanks
    Assign points if this helps

  • Issue with the Posting Date of the Purchase Order.

    Hi All,
    There are fields in BW like SSL1: Time OK, SSL2: Qty OK, SSL3: Time & Qty Ok, SSL4: Days Late (Routines are written to calculate). These fields will indicate whether the delivery against a GR is OK or not with respect to Time, Quantity and the No. of Days..
    But here the issue I am facing is
    If there is only1 delivery/ GR against a single item the calculation in BW are correct - i.e. for a particular PO if there is only one delivery the above fields like SSL1: Time Ok, SSL2: Qty OK will show like the delivery is done within the specified time and everything is OK (in case if it is delivered within the allotted time)
    But if there are multiple deliveries or multiple GR's  posted for one PO item, the calculations are going wrong i.e. even if the delivery is done well within the specified time it is showing the wrong calculations like it is delievered too late. Because in this case the earlier dates are overwritten.
    Can anyone throw me some light on how can I go about solving this issue.
    I am thinking of declaring the Posting Date as the KeyField of the DSO as of now it is a Data field  I also want to know the impact of assigning this as a Keyfield.
    Thanks in advance,
    Prasapbi

    Hi,
    As I understand, you have a DSO based on Purchase Order and your key field is PO and its line item. The problem as you stated will always be there if you have multiple deliveries/GRs created for a single line item because the system will overwrite the entries for same key.
    Problem with adding Posting date as keyfield will be that then your key will be PO-PO Lineitem-Date. When PO will be created, the Posting date will be blank(correct me here if I am wrong), therefore you will have one entry for same PO-line item combination. One without date and other with date, which again would be incorrect. If my assumption about Posting date was wrong, even then your data may not be correct because then you may have many entries with same posting date which again would overwrite each other.
    If there is any direct link between PO line item and number of deliveries that will get created for them, then you can bring that field in DSO as keyfield. But I don't think there is any such field.
    Looking at your report requirement, I would suggest that you make a DSO based on Goods Receipts and then calculate these keyfigures by comparing the dates between GR posting date and PO line item date.
    Else you can change the way your datasource works(if its generic one based on function module). Since your main requirement is to check whether the GR posting date has met your SLA or not, you should fetch all the details only when GR is created and make your key field as PO-PO Line item-GR

  • Issue with calling an external web service using web service proxy

    Hi,
    I've created several web services using JDeveloper succesfully in the past, however I'm getting an issue with one that was working ok previously. Have spoken to the developers of the external web service and they say they haven't changed anything. Also if I test the external web service through soapUI for example it seems fine so it would seem that the issue is with the autogenerated code created by JDeveloper.
    Basically I get the following exception as it sends the request to the external service:
    java.security.PrivilegedActionException: javax.xml.soap.SOAPException: Message send failed: Bad Authentication header format: 'Basic realm="Integration Server" encoding="UTF-8"'
    Expected "," at position 33
    The user name/password appear correct.
    JDev version 10.1.3.42.70
    Any ideas?
    Many Thanks
    Gary
    Edited by: user10916721 on 08-May-2012 09:54

    Found the answer. Basically I had one proxy web service working and one not both linking to a middle tier written in web methods. THe issue actually ended up being with the web methods code in that the one web service (that was not working) was put together using a backwards compatability mode. Once this was switched back then started working again in JDev.

  • Issues with 4.1 Data Upload

    I've got some issues with the new feature 'data upload' in 4.1. I had already made the pages and the data loading and it worked perfect. Now a new column was added to my data loading table and I have to add a new table lookup for this column. Everything is set up like it has to be. I've got 4 table lookups. When I try to upload some data and indicate the table columns to the data, I've always get the error: Failed to retrieve the lookup value. There is no problem when I do a data load where he only have to retrieve one column from the lookup table, when he has to retrieve data from more tables for more columns, I always get the FAILED message. Does anyone know the problem of this situation? I already tried to make a totally new data loading, but this also failed to do the job.

    Hi Ellenvanhees,
    I dont think the number of lookups that you defined is an issue here. If possible try to share more details about your data and tables. The few things that come to my mind are probably your data.
    But if you are able to do one by one lookup without problem then I think your upload is failing due to null values. The current status of data upload feature returns a failed lookup even if a null value was to be uploaded. This is a bug #13582661 and has been fixed for the future release.
    Patrick

  • Issue with setting Planning Data Source

    I keep getting error when I try to set up the data source
    the Oracle database  = orcl
    The data base schema EPMAPLAN
    Password = EPMAPLAN
    when i log on i  to Oracle Database
    Sys as SYSDBA
    Password = Password01
    so what what should be the Database Source Name  = ORCL Correct ????
    Server = win-392h1l307n1 or localhost
    Databse = ORCL
    User Name  = ????????????
    Password  = ?????????
    I try all the  combinations
    Please advise

    Duplicate post - Issue with setting up planning data source
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Has anyone experience iMac memory issues with One Drive (previously Sky Drive) - uses 5GB of my 8GB memory!

    I recently started experiencing memory issues with my iMac due to One Drive using up to 5GB of my iMac memory (total of 8GB). Every time I start up my iMac Activity Monitor shows One Drive uses around 4-5Gb of the internal memory for no reason. Previously Onde Drive seemed to work fine, it's only since about 2 weeks I notice this. It's really annoying as the iMac freezes up regulalrly and I need One Drive to synchronise my files across different devices.
    Has anyone experienced this too and are there any known fixes?

    Yes this is happening to me too, since the last upgrade. Eventually MacOS starts insisting I "Force Quit" all my applications as I have run out of application memory.
    OneDrive doesn't appear in that list because it's not an application... but Activity Monotor clearly shows it's the culprit, taking up 5GB+ of RAM.
    I've raised the issue on Microsoft's forums:
    http://answers.microsoft.com/en-us/onedrive/forum/sdperformance-sdother/onedrive -for-macos-memory-leak/fe04cf60-949f-47b6-b886-10539b5c5cc3?tm=1397636232934

Maybe you are looking for