OEID Big Data Connector Integration

In the oracle big data solution , it shows OEID can source from Oracle Big Data connectors to ingest Big Data for exploration and discovery.
Is there any documentation around as to how to achieve this ?
Can anyone share any information ?

It depends on the connector. Oracle Loader for Hadoop runs on the Hadoop cluster, and does not reside on the Oracle Database. It can reside on a node from where you submit your map reduce nodes - which might or might be a node on the cluster. Oracle Direct Connector for HDFS resides on Oracle Database.

Similar Messages

  • Connection with Hadoop/big data

    Hi,
    how can we connect the HADOOP with the ENDECA STUDIO. Any document please do let me know.
    thanks and regards
    Shashank Nikam

    Hi,
    As far as I know you will need to use Oracle Data Integrator (ODI) or Oracle Big Data connectors
    Some sites:
    Oracle Data Integrator Enterprise Edition 12c | Data Integration | Oracle
    http://docs.oracle.com/cd/E37231_01/doc.20/e36963/concepts.htm#BIGUG107
    http://www.oracle.com/us/products/database/big-data-connectors/overview/index.html

  • Is there any connectors between - OBIEE RPD & Big Data

    Is there any connector between - OBIEE RPD & Big Data? How we will get structured & unstructured data into OBIEE RPD to generate reports?

    Not sure what you want to achieve here... however.. what I read is that you would like to have a way to get unstructured data into structured so you can report on it using OBIEE. (Correct?).
    Depending on the amount of data and the things you need to do to it before you can structure it you can create a map/reduce function and use Hadoop. There is a connector to push the hadoop results into the Oracle database. Then you could use OBIEE to report on the now structured data in the Oracle database.
    As I do not know the details of what you like to achieve,..... this is a solution I would start looking into if I where you. :-)
    Regards,
    Johan Louwers

  • Upcoming Lumira Webinar June 11th on the topic of "Big Data Visualization and Custom Extensions"

    The next webinar in the Lumira series is coming up this week on Wednesday, June 11th, 10:00 AM - 11:00 AM Pacific Standard Time. The day is almost here and if you've already registered, thank you!
    If not, please click here to register.
    Speaker Profile:
    The topic is presented by the Jay Thoden van Velzen, Program Director Global HANA Services/Big Data Services Center of Excellence at SAP!
    Jay has been working in Analytics/Business Intelligence since it was called Decision Support Systems in the late 90s. Currently he is focused on Big Data solutions and how to make the various components of such a solution run smoothly integrated together using the SAP HANA Platform. 
    Abstract:
    Big Data analysis poses unique and new challenges to data visualization, compared to more traditional analytics. Such analysis often includes frequency counts, analysis of relationships in a network, and elements of statistical and predictive modeling. In many cases, traditional visualization techniques of bar- and column charts, pie charts and line graphs are not the most appropriate. We have to avoid the “beautiful hairball” and make it easy for end users to absorb the information through clever use of filtering, transparency and interactivity. We will likely also need to provide more context to go with the visualization than we have been used to in traditional analytics. Moreover, in case of forecasts you need to include any confidence intervals in order not to mislead.
    This means we need more chart types, and often the chart types you need may not exist, nor could the need for such chart types necessarily be foreseen. However, Lumira allows us to design and code our own D3.js visualizations and integrate it into Lumira while providing all the data access methods – including SAP HANA – that it provides out of the box. This means we can develop our visualizations to share the outcomes of Big Data analysis to exactly how we feel it should be presented. During the webinar we will show a number of examples, and specifically the integration of forecasts coming from R into Lumira through a Lumira custom extension.
    We really hope to see you there!
    Cheers!
    Customer Experience Group

    Congrats to Joao and Alex!
     Microsoft Azure Technical Guru - May 2014  
    João Sousa
    Microsoft Azure - Remote Debbuging How To?
    GO: "Clever. Well Explained and written. Thanks! You absolutely deserve the GOLD medal."
    Ed Price: "Fantastic topic and great use of images!"
    Alex Mang
    The Move to the New Azure SQL Database Tiers
    Ed Price: "Great depth and descriptions! Very timely topic! Lots of collaboration on this article from community members!"
    GO: "great article but images are missing"
    Alex Mang
    Separating Insights Data In Visual Studio Online
    Application Insights For Production And Staging Cloud Services
    Ed Price: "Good descriptions and clarity!"
    GO: "great article but images are missing"
    Ed Price, Power BI & SQL Server Customer Program Manager (Blog,
    Small Basic,
    Wiki Ninjas,
    Wiki)
    Answer an interesting question?
    Create a wiki article about it!

  • CC&B and Meter Data Management Integration Usage Requests

    Hello.
    We investigating the following issue: Trying to create a Bill in CC&B with getting usage data from Meter Data Management using Usage Requests engine. Help pages of CC&B describe the big picture of Usage Requests and configuration overview. So this information is not enough for correct and complete adjustments. There is no any information about outbound and inbound XML messages formats.
    CC&B help contains link to the document: Oracle Utilities Customer Care and Billing - Meter Data Management Integration Implementation Guide, but our attempts of searching of this document was ineffectual.
    It's interesting to get any information about Usage Requests from CC&B to Meter Data Management or link to the document mentioned above.
    Thanks in advance!
    P.S. Our current setup of Usage Request gives the next log during bill segment generation (we are using jms queues and xsl transformation of outbound and inbound messages. So MDM sends a response, but something wrong in further processing on CC&B side). As a result we get a bill segment in Error state with remark "Usage Request for Bill Segment is not found.".
    Fragment from weblogic_current.log:
    SYSUSER - 318312-3794-1 2011-01-19 14:40:55,547 [Remote JVM:1 Thread 3] WARN (host.sql.CobolSQLParamMetaData) COBOL set the null indicator to false for SQL bind parameter xWIN-START-DT, but it sent a null value ' '; binding null
    SYSUSER - 318312-3794-1 2011-01-19 14:40:55,554 [Remote JVM:1 Thread 3] INFO (COBOL.CMPBBLLP) Invoking CMPBBILG
    SYSUSER - 318312-3794-1 2011-01-19 14:40:55,554 [Remote JVM:1 Thread 3] INFO (COBOL.CMPBBILG) CMPBBILG validate
    SYSUSER - 318312-3794-1 2011-01-19 14:40:55,572 [Remote JVM:1 Thread 3] INFO (COBOL.CMPBBILG) Get next Acct sa
    SYSUSER - 318312-3794-1 2011-01-19 14:40:55,579 [Remote JVM:1 Thread 3] INFO (COBOL.CMPBBILG) Acct Sa more
    SYSUSER - 318312-3794-1 2011-01-19 14:40:57,394 [Parent Reader:Thread-54] INFO (api.maintenanceObject.MaintenanceObjectLogHelper) (Server Message)
         Category: 11002
         Number: 12150
         Call Sequence:
         Program Name: Usage_CHandler
         Text: Transitioned from Pending to Awaiting Data Sync.
         Description:
         Table: null
         Field: null
    SYSUSER - 318312-3794-1 2011-01-19 14:40:58,089 [Parent Reader:Thread-54] INFO (api.maintenanceObject.MaintenanceObjectLogHelper) (Server Message)
         Category: 11002
         Number: 12150
         Call Sequence:
         Program Name: Usage_CHandler
         Text: Transitioned from Awaiting Data Sync to Send Request.
         Description:
         Table: null
         Field: null
    SYSUSER - 318312-3794-1 2011-01-19 14:41:03,720 [Parent Reader:Thread-54] INFO (support.context.CacheManager) Registering cache 'XAIOptionCache'
    SYSUSER - 318312-3794-1 2011-01-19 14:41:06,920 [Parent Reader:Thread-54] INFO (domain.integration.RealtimeOutboundMessage) sending Realtime Outbound message <?xml version="1.0" encoding="UTF-8"?><SOAP-ENV:Envelope xmlns:SOAP-ENV="urn:schemas-xmlsoap-org:envelope"><SOAP-ENV:Body><CM_UsageCalculation transactionType="READ" dateTimeTagFormat="CdxDateTime"><rate>ERES-1</rate><saId>123456789</saId><usageId>954617571747</usageId><scalarProcessing><billingOption>Y</billingOption><startDateTime>2010-06-09-12.00.00</startDateTime><endDateTime>2010-06-10-12.00.00</endDateTime></scalarProcessing></CM_UsageCalculation></SOAP-ENV:Body></SOAP-ENV:Envelope>
    SYSUSER - 318312-3794-1 2011-01-19 14:41:22,493 [Parent Reader:Thread-54] INFO (domain.integration.RealtimeOutboundMessage) Raw Response from External System is ID:325496.1295437282326.0
    SYSUSER - 318312-3794-1 2011-01-19 14:41:23,966 [Parent Reader:Thread-54] INFO (support.schema.BusinessObjectInfo) BusinessObject C1-NonCyclicalUsgReqOutMsg: Skipping audit calls while performing DEL on entity OutboundMessage_Id(899647019991), as there were no auditable changes
    SYSUSER - 318312-3794-1 2011-01-19 14:41:24,655 [Parent Reader:Thread-54] INFO (api.maintenanceObject.MaintenanceObjectLogHelper) (Server Message)
         Category: 11002
         Number: 12150
         Call Sequence:
         Program Name: Usage_CHandler
         Text: Transitioned from Send Request to Awaiting Bill Determinants.
         Description:
         Table: null
         Field: null
    SYSUSER - 318312-3794-1 2011-01-19 14:41:24,939 [Parent Reader:Thread-54] INFO (api.maintenanceObject.MaintenanceObjectLogHelper) (Server Message)
         Category: 11002
         Number: 12150
         Call Sequence:
         Program Name: Usage_CHandler
         Text: Transitioned from Awaiting Bill Determinants to Bill Determinants Received.
         Description:
         Table: null
         Field: null
    SYSUSER - 318312-3794-1 2011-01-19 14:41:25,892 [Remote JVM:1 Thread 3] WARN (host.sql.CobolSQLParamMetaData) COBOL set the null indicator to false for SQL bind parameter xWIN-START-DT, but it sent a null value ' '; binding null
    SYSUSER - 318312-3794-1 2011-01-19 14:41:25,932 [Remote JVM:1 Thread 3] INFO (COBOL.CMPBBILG) Get next Acct sa
    Edited by: Anton on 19.01.2011 1:56

    Did you find the document Oracle Utilities Customer Care and Billing - Meter Data Management Integration Implementation Guide?
    I searched on eDelivery and Oracle Support, but cannot find the document.

  • WebCenter Content Connector Integration

    Hello,
    I have a problem to test the WCC Connector integration for WebCenter Sites. I had and Error -3 when connector wants to save all assets from WCC.
    I don't have external Ldap I use Weblogic for WCC and I use internal for WCS
    Thanks for yours help
    Michel

    I assume it will require more data to be able to analyse where the problem potentially can be.
    First of all take a look at these two articles that explain what needs to be done. Perhaps, you will find if something is missing:
    http://www.ateam-oracle.com/webcenter-sites-and-webcenter-content-integration/
    WebCenter Sites 11g R1 (11.1.1.8.0) First Look: The WebCenter Content Connector | Function1

  • BIG Data with OBIEE

    What is Big Data? How to handle that in OBIEE?
    Please give me the steps how to implement that.
    Thanks.

    Hi,
    Not sure what you want to achieve here... however.. what I read is that you would like to have a way to get unstructured data into structured so you can report on it using OBIEE. (Correct?).
    Depending on the amount of data and the things you need to do to it before you can structure it you can create a map/reduce function and use Hadoop. There is a connector to push the hadoop results into the Oracle database. Then you could use OBIEE to report on the now structured data in the Oracle database.
    As I do not know the details of what you like to achieve,..... this is a solution I would start looking into if I where you.
    mark as helpful if answered

  • Refresh System Data in Integration Directory PI 7.11

    Hey,
    I have setup a new technical system and a business system in SLD, but now I still can't find it in Integration Directory.
    I knew from the later PI Version, that there was an option to transfer SLD data to integration directory, but now I can't find it
    anymore.
    Can anybody help?
    Thanks!
    Michael

    I have setup a new technical system and a business system in SLD, but now I still can't find it in Integration Directory.
    I knew from the later PI Version, that there was an option to transfer SLD data to integration directory, but now I can't find it
    Tools (or right-click in ID) --> Assign Business System --> follow the wizard...it will assign the business system from SLD to ID

  • I am using the big date calendar template and when I submit it to apple for printing I lose the name of two months. These names are not text boxes. I see the names when I send it in but something happens during the transmission to apple. It was suggested

    I am using the big date calendar template in iPhoto. I am on Lion 10.7.2, macbook air. The names of the months are on each calendar page but something happens when I send the data to Apple. The names are part of the template. They are not text boxes. I lose two names on the calendar after it is sent to Apple. Apple suggested I make a pdf file of my calendar before sending it in and check to make sure every name shows. I did this with a calendar I just sent in. The calendar was correct. All names of the months were showing. After sending the data two month names disappeard because when it arrived by mail, it was incorrect. Apple looked at my calendar via a pdf file and it was incorrect.  This is second time this has happened. I called Apple and they had me delete several folders in the Library folder, some preferences and do a complete reinstall of iPhoto.  I have not yet remade the defective calendar. I am wondering if anyone else has had this problem?
    kathy

    Control-click on the background of the view all pages window and select "Preview Calendar" from the contextual menu.
    You can also save the pdf as a file to compare to the printed calendar.  If the two names are visible in the pdf file then the printed copy should show them.  Contact Apple for a refund.  Apple Print Products - Apple Store (U.S.)

  • Please Help PI Data Dependent Integration Builder Authorizations NOT Workng

    Dear Friends / Experts,
    I had spend many days and explored all Weblog  and links on this website and implemented all the steps required to acheive Data Dependent Integration Builder Security and I am not successful so far. I am just giving up now - Please Help Me ---
    As I said, I already read all the important Forum Links and SAP Web links and Followed Each and Every Step - service.sap.com/instguidesNW04 ® Installation ® SAP XI
    Security Requirement - Data Dependent/Object Level Authorizations in XI / PI
    In distributed teams or in a shared PI environment it might be necessary to limit authorization for a developer or a group of developers to only one Software Component or objects within a Software Component or to specific Configuration Objects.
    Our Environment - PI 7.0 SP 16
    Created a new role in the Integration Builder Design
    u2013Add Object Types of any Software Component and Namespace
    - Enable usage of Integration Builder roles in Exchange Profile
    Integration Builder u2013Integration Builder RepositoryParameter com.sap.aii.util.server.auth.activation to true
    Assign users to the newly created Integration Builder roles
    u2013Create dummy roles in Web AS ABAP, these roles are then available as groups in Web AS Java
    u2013Assign users to these roles
    u2013Assign the Integration Builder roles to the above groups in Web AS Java
    u2013Assign unrestricted roles to Super Users
    Please help - How to validate whether Data Dependent Authorizations are Activated?
    I am working with XI Developers and Basis Team and we did updated all the Required Exchange profile parameters.
    Per this Document - User Authorizations in Integration Builder Tools - Do we need to update the server.lockauth.activation in Exchange Profile. When We updated, It removed Edit Access from all XI Developers in PI
    In both the Integration Repository and the Integration Directory, you can define more detailed authorizations that restrict access to design and configuration objects.
    In both tools, you define such authorizations by choosing Tools ® User Roles from the menu bar. The authorization for this menu option is provided by role SAP_XI_ADMINISTRATOR_J2EE. Of course, this role should only be granted to a very restricted number of administrators. To activate these more detailed authorizations, you must set exchange profile parameter com.sap.aii.ib.server.lockauth.activation to true.
    The access authorizations themselves can be defined at the object-type level only (possibly restricted by a selection path), where you can specify each access action either individually as Create, Modify, or Delete for each object type, or as an overall access granting all three access actions.
    http://help.sap.com/saphelp_nw04/helpdata/en/f7/c2953fc405330ee10000000a114084/frameset.htm
    I was able to control display and maintain access from ABAP Roles, but completely failed to implement Integration Builder Security?
    Are there any ways to check Whether Data Dependent authorization or J2EE Authorizations are activated?
    Thanks a lot
    Satish

    Hello,
    so to give you status of our issue.
    We were able to export missing business component .
    But we also exported some interfaces after that and we had some return code 8, due  to objects still present in change list on quality system (seems after previous failed transports , the change list was not cleared completley...).
    So now we have checked that no objects is present in the change list of quality system and we plan to export again our devs on quality system.
    Hope after that no more return code 8 during imports and all devs transported correctly on quality system.
    Also recommending to read that, which is pretty good.
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/7078566c-72e0-2e10-2b8a-e10fcf8e1a3d?overridelayout=t…
    Thanks all,
    S.N

  • Problems with HowTo "Creation of BI Master Data in Integrated Planning...)

    Hi,
    one of our customers need a planning application in which he can create master data for different attributes, insert some key figures for this attributes and changes already planned key figured.
    For the last two point I've already found a working solution, the first point (creating of masterdata in web template) is still a problem. After a little search, I found a howto-guide "Creation of BI Master Data in Integrated Planning (IP) through Web Layouts", in which the necessary steps are described, how to do this (please see following link: http://www.sdn.sap.com/irj/scn/index?rid=/library/uuid/607193d5-cdd5-2b10-c699-8ff04c3124f6)
    Does anyone of you know this guide and was able to build a working solution?
    I have different questions, maybe you could help me:
    - what kind of variable is he using?
    - do I need a DataProvider in Web Template?
    - how do I get the connection to the used variable in the button group?
    Thanks for your help.
    Regards
    Tim

    Hi Timo,
    you require two howtos to implement the solution.
    First, please take a look at the first document:
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/10d2b273-0e12-2c10-fab3-a34bde559f92?QuickLink=index&…
    In chapter 3 you see the prerequisites (note 1101726 and notes 1384495, 1387004 and 136772). By this, you change the system behaviour, not to check if the new inserted master data is valid (it can't be valid, because the entries are new!)
    And after this please implement the solution from the how to.
    Furthermore you can take a look at this how to, which might help you too:
    http://www.sdn.sap.com/irj/scn/go/portal/prtroot/docs/library/uuid/607193d5-cdd5-2b10-c699-8ff04c3124f6?QuickLink=index&…
    Even if my first post is a bit old, my customer still uses this solution, at the moment we are on 7.3 sp06, so it is working in newer releases too.
    Hope this helps.
    Regards
    Tim

  • Best strategy to upload a big data file and then to insert in db the content

    Hi,
    Here's our requirement. We have a web business application developed on JSF 1.2, JE66, WebLogic for application server, and Oracle for the data back end tier. We need to upload big data files (80 to 100 Mb) from a web page, and persist the content in database tables.
    What's the best way in terms of performance to implement this use case ? Once the file is uploaded on server a command button is available on the web page  to trigger a JSF controller action in order to save data in database.
    Actually we plan to keep in memory the content of the http request, and call insert line on each line of the file. But i think it's bad and not scalable.
    Is is better to write the file on server's disk and then use multiple threads to send the lines to the database ? How to use multi threading in a JSF managed bean ?
    Thanks

    In addition, LoadFromFile is overloaded to handle both BLOB and CLOB:
    PROCEDURE LOADFROMFILE
    Argument Name                  Type                    In/Out Default?
    DEST_LOB                       BLOB                    IN/OUT
    SRC_LOB                        BINARY FILE LOB         IN
    AMOUNT                         NUMBER(38)              IN
    DEST_OFFSET                    NUMBER(38)              IN     DEFAULT
    SRC_OFFSET                     NUMBER(38)              IN     DEFAULT
    <BR>
    PROCEDURE LOADFROMFILE
    Argument Name                  Type                    In/Out Default?
    DEST_LOB                       CLOB                    IN/OUT
    SRC_LOB                        BINARY FILE LOB         IN
    AMOUNT                         NUMBER(38)              IN
    DEST_OFFSET                    NUMBER(38)              IN     DEFAULT
    SRC_OFFSET                     NUMBER(38)              IN     DEFAULT

  • Schema Mapping in Data Service Integrator

    Hi,
    I'm just examining some schema mapping programs like BizTalk, Altova and IBM Rational Data Architect, and now found Oracle Data Service Integrator which might be a similiar tool. Actually I found out about Data Service Integrator via BeQ AquaLogic, which probably had been some schema mapping tool before Oracle acquired it and now offers it as Data Service Integrator.
    My question is, whether Oracle Data Service Integrator is really applicable for schema mapping/matching, such as creating mappings between xml, csv/flat files or database schemas. I already downloaded it and tried it out, but had troubles creating a map. According to a tutorial you create a physical data service if you wanna do something like mapping, but after I did this there was only a "map" with a source schema. There was no way to add a target schema and map it with the source schema.
    So can I create mappings in Data Service Integrator or are there other products which would be more convenient (Oracle Warehourse Builder for instance)? If so, does anyone know whether there is a good tutorial how to map simple schemas such as xml files in Oracle Data Service Integrator?
    Thank you in advance.

    After you create your physical data services, create a logical data service using your target schema as the 'return type'. Then add functions and use the xquery mapper to map your physical data services (csv, database, xml, web service etc) to your target schema. You can also use logical data services as the input to a logical data service.

  • Attempt to fetch cache data from Integration Directory failed

    HI,
    while checking cache connectivity testing: status is
         green:   Integration Repository     
         green:    Integration Directory     
              green: Integration Server - JAVA     
                  red:Adapter Engine af.axd.aipid     
               yello:Integration Server - ABAP
    Jun 30, 2007 1:16:08 PM - Cache notification from Integration Directory received successfully
    Attempt to fetch cache data from Integration Directory failed; cache could not be updated
    [Fetch Data]: Unable to find an associated SLD element (source element: SAP_XIIntegrationServer, [CreationClassName, SAP_XIIntegrationServer, string, Name, is.00.aipid, string], target element type: SAP_BusinessSystem)
    [Data Evaluation]: GlobalError
    what to do?
    and there is nothing under integration server and integration engine but there is an green status under Non-Central Adapter Engines > from this i am doing send messeage testing fro xi to bi ,
    send message to: http://aibid:8000/sap/xi/engine?type=entry
    payload:
    <?xml version="1.0" encoding="utf-8"?>
    <ns1:MI_VCNdatatoBI
    xmlns:ns1="http://bi.sap.com"
    xmlns:xsi="http://www.w3.org/2001/XMLSchemainstance">
    <DATA>
    <item>
    </BIC/ZG_CWW010>1000<//BIC/ZG_CWW010>
    </BIC/ZVKY_CHK>1<//BIC/ZVKY_CHK>
    </item>
    </DATA>
    </ns1:MI_VCNdatatoBI>
    i can sent message from there (component monitoring > Non-Central Adapter Engines) but unable to get it at message monitoring and at BI side.
    dushyant.

    thanks,
    but i have adepter type XI
    and i am folowing step of this lonk and there is no need to create fild adepter type according to that and almost done but while sending message through config. monitor in RWB it goes but not coming in mess monitoring and at bi side
    see 4.5 > 3 and 4 topic and 4.6 > 3,4,5
    https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/f027dde5-e16e-2910-97a4-f231046429f2
    now what to do?
    dushyant,

  • Best practices for administering Oracle Big Data Appliance

    -        Best practices as part of administration of Oracle Big Data Infrastructure
    -        How do we lock down max space usage per project
    Eg: Project team A can have a max limit of 10 TB space allocated
    -        Restricting roles, access ( Read, Write), place holder for common shared artifacts
    -        Template/procedure for code migration across dev,qa and prod environments etc

    Your data is bigger than I run, but what I have done in the past is to restrict their accounts to a separate datafile and limit its size to the max that I want for them to use: create objects restricted to accommodate the location.

Maybe you are looking for

  • Acrobat X - Read this Page Only not working

    Hello, I have Acrobat Pro X, and need to make accessible PDFs from InDesign. Before I upgraded to Acrobat X, I used to have Acrobat read the text to me to help check the reading order. But in Acrobat X, the "Read this page only" selection does not wo

  • How do I get songs I purchased on a different computer to my IPOD?

    I mainly use my home computer when downloading itunes to my ipod. I have my office computer set up as an authorized computer also. I have downloaded songs into my office computer but they will not show up on my ipod. How can I get them on there? I am

  • How to replicate a customer from CRM to ERP?

    Hi you all! Can anybody help me!? A allready made the settings to replicate the ERP clients to CRM customers, but now I want that CRM be the leading system for customers creation, and I don´t know how to do the replication! Help Please

  • Dynamic Memory is not working all the time

    We are in the process off moving our 2008R2 VM's from the 2008R2 HyperV servers to new Server 2012R2 Hosts. We shut down the VM's copy the files and VHD's to the new CSV's en import the VM in the Hyperv Manager. Then we make them high available in th

  • Business Catalyst licensing verification failed

    The Business Catalyst add-on for Dreamweaver says 'licensing verification failed' File sync in on We've tried downloading it manually and installing it through Extension Manager, but we get the same.