Libraries need for ADF Data visualization

Hi all,
I want to ask list of libraries which needed by ADF Data visualization.
Thanks
Hadi Wijaya

Hi,
since you ask this question on the JDeveloper 10.1.3 forum, just be aware that DVT components work on JSF 1.2 only and not in JDeveloper 10.1.3
Frank

Similar Messages

  • ADF Data Visualization

    Hi all,
    a simple question how can i use adf data visualization
    i can not find them in component palette?
    do i need to install something?
    if there are examples please provide me with it
    than you 4 everybody

    The Question in other way :
    How can i Drag and drop my view object in my jsf page as a graph ( Pie Chart,Gannt Chart,.....)
    Thank you all,

  • ADF Data Visualizations group not showing up in Component Palette

    I apologize for my ignorance, but I cannot find how to solve this problem. I am trying to use the data visualization components such as graphs but my component palette does not show this. It does show ADF Faces, JSF, and some others, but not this. I tried adding more libraries to my project's properties, but it does not show up. I am using:
    JDeveloper
    Studio Edition Version 11.1.1.2.0
    Build JDEVADF_11.1.1.2.0_GENERIC_091029.2229.5536
    Any ideas?
    thanks,
    Kerry

    Go into project properties -> JSP libraries and make sure you add the data visualization tag library.

  • Using the Oracle ADF Data Visualization Components in a standard JSF enviro

    Hi,
    I have asked this question on the JDeveloper and ADF forum, haven't gotten a response yet:
    Using the Data Visualization Components from ADF in a JBoss 7.1.1
    .. but I would like to ask the question here as well if anyone have used the data visualization components from Oracle ADF Essentials .. the free version ... using the graph components in general, but I am looking at specifically using the gantt chart component in particular ... in a standard JSF environment ( e.g. Tomcat + Mojarra, or JBoss with the built-in Mojarra implementation ) ??
    I know there are other JSF charting components out there ...like the ones from PrimeFaces, but the gantt chart from ADF looks more ... well .. professional.

    People are not prone to answer questions that are not really a problem. You want to know something about a free product, install it and see for yourself. In any case spreading your question around is not going to help you, the people who know about ADF are still going to be looking in the ADF forum. If you don't get an answer there, what chance do you think you have that you'll get one here where it is off-topic?
    I'll close with a word of advice: you generally pay for quality. If you want professional looking charts, you'll probably have to look at commercial products.

  • Input needed for writing data back into BW/ECC data

    Hello Everyone,
    Can anyone please let me know an example or process you have used for writing data back into BW (DSO/Cubes) or ECC systems.
    We do not have integrated planning in our current system configuration currently.
    So, any sample code for using any of the BAPI/RFC for writing data would be appreciated.
    Also, I am trying to find out if there is any way to schedule the models developed in VC 7.1 in background to automate certain type of data processing which goes across multiple systems.
    Any help would be appreciated.
    Thanks.

    Hello,
    Can anyone please help me out on this one....
    I am aware of few BAPI's such as RSDRI_ODSO_INSERT_RFC but I am not aware of what action has to be used to transfer the data from the table within VC to this BAPI and how to define the parameters as the one's available for the BAPI I mentioned do not fit into the data I have in the table within VC.
    The following are the columns I have in the table within VC,
    1. GL Account
    2. Credit Amount
    3. Debit Amount
    4. Sales Amount
    I have defined the DSO with the same where G/L Account is the keyfield and the rest being data fields.
    Any help would be really appreciated.
    Thanks.

  • Help for ADF data binding with MySQL BLOB type.

    Hi,
    I have a blob column called photo for table person in my MySQL database, with ADF, I want to bind this column with a JUImage control, but when I drag the column attribute from the data control palete as a JUImage, an error dialog appears, it says:
    Control cannot be bound to that Attribute.
    This control should be bound to an Attribute whose Java type is one of the following
    oracle.ord.im.OrdImageDomain
    oracle.jbo.domain.BlobDomain
    oracle.jbo.domain.Raw
    What should I do ?
    Thanks.

    Hi,
    Are you using BC4J as the persistence layer? If you use your own persistence (e.g. a bean) then you need to make sure the datatype of the atribute is defined as one of the three mentioned in your question.
    Frank

  • Help needed! ADF Data Control: Navigation Buttons + Table  issue.

    Hey all!
    We are working on a new application using ADF JSF+EJB 3.0 stack and are experiencing a problem with a simple feature we need to implement.
    Scenario:
    - Creating a data control for a facade session bean and drag a List returned by one of its methods as table;
    - Dragging its corresponding navigation buttons (first, previous, next, last) from the data control onto the page.
    Issue:
    As we press those buttons, we see the page being submitted but but nothing changes in the table...It keep showing the same records all the time (buit-in range navigation works fine though). However, when we manage to drop the collection as a form (with nav buttons included), everything works fine.
    We have inspected pagedefs for both cases and could not identify the difference.
    Any help would be much appreciated as we need to address the issue asap..
    Thanks in advance,
    Gustavo
    PS: Sorry for my English! ;)

    Hey, i have the solve to my problem.
    In the definition page of the table page i change the propert Refresh (ifNeeded is his default value) to always of the iterator who im using to the table.But i use Toplink, then i had to change the property "Refresh Identity Map Results" to "enabled", that option is in the options of the named queries on queries in the toplinkmap of the current table who im using to my page.
    i hope this will be util for you.
    That's all
    Dany

  • ........ Expert  suggestion needed for Loading Data in DB

    Hi all
    The current requirement we have is to Load data from Excel File to oracle Databases.
    The Excel file has around 30 Sheets , each corresponds to a table, ( means 30 tables in the database) .
    Currently we are using sqlldr commands to load data from CSV files to the Oracle Database. The only problem is that sometime the DBA has to go out or is busy workign on somthing else, so May i kindy get some expert suggestions on how to automate this DataLoading Task.
    Somebody has suggested to My Manager ( who is not aware of oracle and IT at all) that data can be loaded via ODBC. It is suggested to him that all we need to do is to place the CSV files on the Server at a particular folder and Oracle will do the rest.
    I am not that proficient in Data loading methods.. so may i know any technique which will simplify/automate the Task.
    I mean how can we automate so that every time , the the sql loader scripts run at command promt. I think data base ( Oracle ) as nothing to do with command promt i feel. isnt it ..
    Kindly have ur expert./experienced suggestions.
    I would be highly grateful to all
    Regards

    To automate sqlldr scripts, you would usually write a OS script file that will periodically looks for files in a given directory and run the sqlldr commands:
    - on Windows: .bat or .wsh files to be scheduled with "at" command or windows scheduled tasks tool
    - on Unix: a script shell in the crontab.
    Much more complicate but without any OS script:
    - write a PL/SQL code to read, parse the file to be loaded using UTL_FILE
    package
    - this PL/SQL code must also generate INSERT statements and process errors ...
    - this PL/SQL code can be scheduled with DBMS_JOB package to run a periodic intervals.
    Message was edited by:
    Pierre Forstmann

  • Time needed for automated data transfer - hypothetic

    An EBS client with interest in maintaining integration in their large customer base backoffice and with CRMOD has had past challenges in terms of the overall size of the data they wish to integration and the time it takes to do so. As such, the client wishes to understand from a data size (X GIGS) and time (how much time) it might take to for instance synchronize their data. Since they would not quantify their data size, I can only ask... how much data be moved and how fast. This is urgent. Any experienced help here using strategies such as batch integration leveraging web services, cast iron systems or PIP is of help here!

    Disclaimer
    The Author of this posting offers the information contained within this posting without consideration and with the reader's understanding that there's no implied or expressed suitability or fitness for any purpose. Information provided is for informational purposes only and should not be construed as rendering professional advice of any kind. Usage of this posting's information is solely at reader's own risk.
    Liability Disclaimer
    In no event shall Author be liable for any damages whatsoever (including, without limitation, damages for loss of use, data or profit) arising out of the use or inability to use the posting's information even if Author has been advised of the possibility of such damage.
    Posting
    "And the solution was using an FTP client which was able to send files using multiple parallel threads."
    Yep, although that assumes you have multiple files (for which it can work very well).
    For a single large file, a variation of what Milan described, is using a utility that can slice and dice (and reassemble) a single file across multiple TCP flows.
    "Another possibility would be using some WAN accelerators (like Riverbeds, e.g.) which would sent "local ACKs" to the clients on both sides and improve the efficiency of the WAN file transfer."
    Yep, again, and such products can even "transfer" faster than the line rate.  The latter is possible because they generally compress in-line and also use local caching (negating the need to send some data). One of their issues, their transfer rates can be very variable and inconsistent.

  • Query needed for Cumulative data

    HI Friends,
    I need output like this.
    Frequency     Percent (%) Cumulative
    Frequency     Cumulative Percent
    4468     0.91     4468     0.91
    21092     4.31     25560     5.23
    57818     11.82     83378     17.05
    6274     1.28     89652     18.33
    I am using Oracle 9i.
    My output data like this and I need to write the query for 3 columns (Percent ,Cumulative frequency and Cumulative percent)
    1:The formula for Percent column data is (Frequency/Sum of cumulative frequency)*100
    2:The formula for Cumulative Frequency column data is (Cumulative of Frequency column data)
    3:The formula for Cumulative Percent column data is (Cumulative of Percent column data)
    What should be the analytic function and how to write the query.
    Thanks,
    Lony

    Hi Friends,
    I need output like this.
    Frequency Percent (%) Cumulative Frequency Cumulative Percent
    4468 0.91 4468 0.91
    21092 4.31 25560 5.23
    57818 11.82 83378 17.05
    6274 1.28 89652 18.33
    I am using Oracle 9i.
    My output data like this and I need to write the query for 3 columns (Percent ,Cumulative frequency and Cumulative percent)
    1:The formula for Frequency column data is sum of (dd+cc+mc_cc_mc).
    1:The formula for Percent column data is (Frequency/Sum of cumulative frequency)*100
    2:The formula for Cumulative Frequency column data is (Cumulative of Frequency column data)
    3:The formula for Cumulative Percent column data is (Cumulative of Percent column data)
    What should be the analytic function and how to write the query.Please find the sample data and table script.
    CREATE TABLE all_lony (
    campno varchar2(20),
    dd INTEGER,
    cc INTEGER,
    mc INTEGER,
    cc_mc INTEGER
    insert into all_lony (campno,dd,cc,mc,cc_mc)
    values(36,156,1320,445,2547);
    insert into all_lony (campno,dd,cc,mc,cc_mc)
    values(40,233,19711,263,885);
    =============
    Please find my query below
    SELECT campno
    || ','
    || dm
    || ','
    || cc
    || ','
    || mc
    || ','
    || cc_mc
    || ','
    || frequency
    || ','
    || per
    ||','
    ||cumulative_fr
    ||','
    || SUM (per) OVER (ORDER BY per ROWS UNBOUNDED PRECEDING)
    FROM (SELECT q3.campno campno, q3.dm, q3.cc, q3.mc, q3.cc_mc,
    q3.frequency, q3.cumulative_fr,
    (q3.Frequency / SUM (q3.cumulative_fr)) * 100 per
    FROM (SELECT q2.campno campno, SUM (q2.dm) dm, SUM (q2.cc) cc,
    SUM (q2.mc) mc, SUM (q2.cc_mc) cc_mc,
    (SUM ( NVL (q2.dm, 0)
    + NVL (q2.cc, 0)
    + NVL (q2.mc, 0)
    + NVL (q2.cc_mc, 0)
    ) frequency,
    SUM (SUM ( NVL (q2.dm, 0)
    + NVL (q2.cc, 0)
    + NVL (q2.mc, 0)
    + NVL (q2.cc_mc, 0)
    ) OVER (ORDER BY SUM ( NVL (q2.dm, 0)
    + NVL (q2.cc, 0)
    + NVL (q2.mc, 0)
    + NVL (q2.cc_mc,0)
    ) ROWS UNBOUNDED PRECEDING)
    cumulative_fr
    from all_lony
    q1 )q2
    GROUP BY q3.campno) q3
    GROUP BY campno, dm, cc, mc,cc_mc, frequency,cumulative_fr)
    Can anybody just verify the query and let me know.

  • Driver needed for PCI Data Acquisition and Signal Processing Controller of HP 15- r247tu

    Install the Windows 10 Intel IMEI Intel Management Engine Interface (MEI) Driverhttp://ftp.hp.com/pub/softpaq/sp69501-70000/sp69593.exe The latest version of the Intel chipset driverhttp://ftp.hp.com/pub/softpaq/sp69501-70000/sp69575.exe All Windows 10 drivers available from HP are at the following link.http://support.hp.com/us-en/drivers/selfservice/HP-15-Notebook-PC-series/7486447/model/7693973#Z7_3054ICK0KGTE30AQO5O3KA30R1

    Hi: Try the chipset driver directly from Intel...the second file is the auto install file. https://downloadcenter.intel.com/download/20775/Intel-Chipset-Device-Software-INF-Update-Utility- If that driver does not install the PCI Data Acquisition and Signal Processing Controller, which the chipset driver should install, download the first file listed above (zip file). Unzip and extract the contents to its folder. Go to the device manager and click on the above device needing the driver.  Click on the driver tab. Click on Update driver.  Select the Browse my computer for driver software option, and browse to the unzipped driver folder. Make sure the Include Subfolders box is checked, and see if the driver installs that way. Then reboot.

  • Advice needed for using date type in leave quota generation.

    Requirement:   Seniority for quota generation of long service leave must be based on date "02" in IT0041.
    Although all the relevant config has been done using the date type 02, the quota is still being generated based on date type 01. 
    My Question:  What is required to make this scenario work?
    Thanks
    Maureen.

    Validity/Deduction
    Validity:  Valid From = Date type 02 -1Days; Valid To =  Date Type 02 999Months
    Deduction: From = Start of Val +1Day, To =  End of Val +999 Months
    Base Entitlement:
    Rule: 002
    Sequence 001; Seniority = 0 - 5 years; Constant= 5 days; Related to = Accrual period
    Sequence 002; Seniority = 5 - 999 years; Constant=10 days; Related to = Accrual period
    Selection Rules:
    Applicability:  default values
    Accrual Period: Rel to date type = 02; Length = 60 Months
    Base Entitlement: Rule = 002; Key for seniority = For exact day
    Accrual Entitlement: No pro-rata calc
    Transfer time:  Upon accrual
    All fields not mentioned are standard defaults.
    Employee:
    IT0041
    Date Type 01 = 01/09/2005
    Date Type 02 = 06/04/1995
    IT0000
    Current position:  Begda: 01/02/2009 - 31/12/9999
    Previous position:  Begda: 01/03/2008 = 31/01/2009
    Testing by running Time Evaluation till 31/09/2012
    Result:
    5 days accrued on 05/04/2010
    IT2006:  Deduction from = 06/04/2009 To 31/12/9999;  Validity from = 05/04/2009 To 31/12/9999
    Please note I've used different dates in an attempt to clearly illustrate the problem. Hope it doesn't serve to confuse.
    Thanks for the help.

  • What chords do i need for a data migration from macbook 2008 to  macbook air 2012

    My 2008 macbook has a firewire port and my new macbook air 2012 has a thunderbolt port. I have a thunderbolt cable but i can't find an adapter to firewire or the firewire cord? Should I return the thunderbolt cable and surrender the 100 bucks for one-on-one or go through the tedious process of using my external hard drive as the 'messenger boy'?

    There's all sorts of ThunderBolt to FireWire adapters out there. Here's one on Amazon. http://www.amazon.com/Apple-Thunderbolt-to-FireWire-Adapter/dp/B008RXYOKY/ref=sr _1_11?ie=UTF8&qid=1346277186&sr=8-11&keywords=thunderbolt+firewire
    You'd also need a FireWire 800 to 400 adapter. http://www.amazon.com/elago-FireWire-400-800-Adapter/dp/B002TF0ZDQ/ref=sr_1_8?ie =UTF8&qid=1346277133&sr=8-8&keywords=thunderbolt+to+firewire

  • What's the need for capitalization date and 1st acq. posting date?

    hi guys,
    i was wondering what the system would do with capitalization date and first acquisition posting date ??
    and what will be the problem if i entered an asset value date (in the acquisition transaction) that's earlier than the capitalization date entered in the asset master record ?!

    In practice all assets are not put to use right from the date of purchase.Hence the concept of different dates arises.  One being the acquisition date and the other Capitalization date.
    I would buy an asset today but start using it from a later date.  Depreciation has to be calculated from the date of capitalization and not acquisition. But still for record purpose date of purchase would be important.  Hence both the dates are maintained.
    Hope this has clarified your doubts.
    Thanks
    GU

  • ADF Data visualization Gantt chart displaying 10 columns only

    hi
    i am using Gantt chart and it is only displaying 10 columns from the view object the 11 column is not visible
    how can i solve this??
    also these columns are taking so much space, and i want to make their width smaller, but when i change the style --> box --> width, the column alone change its width but without the header.
    thanks

    People are not prone to answer questions that are not really a problem. You want to know something about a free product, install it and see for yourself. In any case spreading your question around is not going to help you, the people who know about ADF are still going to be looking in the ADF forum. If you don't get an answer there, what chance do you think you have that you'll get one here where it is off-topic?
    I'll close with a word of advice: you generally pay for quality. If you want professional looking charts, you'll probably have to look at commercial products.

Maybe you are looking for

  • My iWeb problem stumped the Genius Bar

    Hello all , first post , apologies in advance for its length . I got a problem thats really frustrating me . The Genius man in the apple shop couldn't help , he'd never seen this problem before so he suggested I try here . Here's the thing . I've mad

  • Acrobat X Pro to XI Standard Upgrade Path

    Is there a way to upgrade my Adobe Acrobat X Pro to Adobe Acrobat XI Standard? I find I really don't need the features of the Pro version and would like to save money upgrading to XI.

  • Problem  connection iPad to net gear wn2000rpt

    All appears to be connected, lights are all green but does connect to Internet

  • How can I undisable my apple id?

    How can I undisable my apple id?

  • Will only export as a Self-Contained file

    I am trying to export my project as a QuickTime movie that's not self-contained but even though I have the box unchecked it still writes all the video / audio to the file (making it GBs where it should be KBs). How do I fix this without reinstalling