What is SAP recommended approach to create multiple Sender Idoc_AAE adapters channels?

Hi All,
We have more than 100 interfaces which receives Idocs from multiple SAP ECC Systems. What is the best approach to use Idoc_AAE sender adapter with multiple communication channels for One system and for multiple SAP systems as well.
Q) How many communication channels we can assign or bind to one Resource Adapter(Assume InboundRA)?
Q) What is performance impact if we create more Resource Adapter to Connect multiple SAP System?
Please share your views to create multiple Idoc_AAE sender communication channels to connect one SAP ECC Systems?
Regards,
Sreeramulu Konjeti.

Hi Vertika - It should work without changing the occurrence as it's on the sender side...
In your mapping, you would receive all the IDOC's though the occurrence is 0..1...
Just simply paste the payload in the test tab and see.. it should not throw any error..

Similar Messages

  • What is the recommended approach to create XML input for forms service

    Context:
        As part of our development project, we need to retrieve data (eg. Policy details) from different modules and using this data we need to generate PDF document using a particular template (what we have uploaded to live cycle)
    Solution we have tried:
    We have created Java classes from XSD (for a particular template) using JAXB complier. Using these Java classes we instantiated all the parameters using data from different modules and got XML after marshalling. We send this XML as a byte array to forms service to get PDF. No issues in this context.
    After displaying the PDF, under writer will modify some of the data and submit back to the application. Application again uses forms service to retrieve XML back as a byte array. This is where we had issues. We tried un-marshalling using same JAXB classes but we are loosing data for all the rich text fields.
    Can you please guide us to handle this issue ASAP, because we are completely stuck in the development stage. We really appreciate if some one can respond to this issue.
    Thank you
    Srini P

    Context:
        As part of our development project, we need to retrieve data (eg. Policy details) from different modules and using this data we need to generate PDF document using a particular template (what we have uploaded to live cycle)
    Solution we have tried:
    We have created Java classes from XSD (for a particular template) using JAXB complier. Using these Java classes we instantiated all the parameters using data from different modules and got XML after marshalling. We send this XML as a byte array to forms service to get PDF. No issues in this context.
    After displaying the PDF, under writer will modify some of the data and submit back to the application. Application again uses forms service to retrieve XML back as a byte array. This is where we had issues. We tried un-marshalling using same JAXB classes but we are loosing data for all the rich text fields.
    Can you please guide us to handle this issue ASAP, because we are completely stuck in the development stage. We really appreciate if some one can respond to this issue.
    Thank you
    Srini P

  • What is the best approach to handle multiple FK with single table.

    If two tables are joined with each other with more than one ways, for example
    MAIN table is (col1, col2,....coln, person_creator_id, person_modifier_id)
    PERSON table is (person_id, name, address,........ phone) etc
    At database level PERSON_CREATOR_FK and PERSON_MODIFIER_FK are defined.
    Objective is to create a report that shows
    col1, col2...coln, person creator name, person modifier name
    If above two objects are imported with FKs in a EUL and discoverer plus is used to create above report. On first inclusion of person name discoverer plus will ask you to pick the join (provided the checkbox to disable this feature is not checked). Once you pick 'person creator' join it will never allow you to pick person modifier name.
    One solution is two create a custom folder with query like
    select col1, col2,...coln,
    pc.name, pc.address,.... pc.phone
    pm.name, pm.address,.... pm.phone
    from main m,
    person pc,
    person pm
    where m.person_id_creator = pc.person_id
    and m.person_id_modifier = pm.person_id
    Second solution is to import the PERSON folder twice in EUL (optionally named one as perosn_creator and other as person_modifier) and manually define one join per table. i.e. join MAIN with PERSON_CREATOR on person_creator_fk and join MAIN with PERSON_MODIFIER table using person_modifier_fk.
    Now discoverer plus will let you drag Name from each person folder without needing to resolve multiple joins.
    Question is, what approach is better OR is there a better way?
    With solution 1 you will not be able to use functions on folder items.
    With solution 2 there is a EUL design overhead of including same object multiple times and then manually defining all join (or deleting unwanted joins), and this could be a problem when you have person_modifier and person_creator in nearly all tables. It could be more complicated if person table is further linked other tables and users want to see that information too. (for instance, if person address is stored in LOCATION table joined with location_id and user want to see both creator address and modifier address....now you will have to create multiple LOCATION folders).
    A third solution could be to register a function in discoverer that return person name when person_id is passed. This will work perfectly for above requirement but a down side is the report will run slower if they need filters on person names (then function will be used in where clause). Also, this solution is very specific to above scenario, it will not work if you want the report developer the freedom to pick any attribute from person table (lets say, person table contain 50 attributes then its not a good idea to register 50 functions).
    Any comments/suggestion will be appreciated.
    thanks

    Hi
    In a roundabout way you have really answered your own question :-)
    In my opinion, the best approach, although by all means not the only approach - see below) would be to have the object loaded as two folders with one join going to the first folder and the second join to the other folder. You would of course name the folders appropriately.
    Here's a workflow that I use all of the time and one that I teach when I'm giving Discoverer Administrator training. It might help you:
    1. Bring in the PERSON folder to begin with
    2. Make all necessary adjustments to bring it up to deployment standard. These adjustments would be: folder name (E.g PERSON_CREATOR), item names, item placement, default positions, default aggregation and so on.
    3. Create or assign the required lists of values
    4. Create any required calculations
    5. Create any required conditions
    6. Create the first join from this folder to MAIN.
    7. Click on the heading for the folder and press CTRL-C.
    8. Click on the heading for the business area and press CTRL-V. A second copy of the folder, complete with all of the adjustments you made earlier will be inserted into the business area.
    Note: joins are not copied, everything else is.
    9. Rename this folder to say PERSON_MODIFIED
    10. Rename the items as appropriate
    11. Add a join from this folder to MAIN - you're done
    Other ideas that I have used and work well would be to use a database view or create a complex folder. Either will work, In both cases you would need to join on some other column other than the ones you referred earlier.
    I hope this helps
    Best wishes
    Michael

  • What is the recommended way to do multiple channel, single point sampling for control with an NI PCI-6255 in RLP?

    Hello,
    I am writing a driver for the M-series NI PCI-6255 for QNX. I have downloaded the MHDDK and have all the examples working. I have also enhanced the examples to do interrupt handling (e.g. on AI_FIFO interrupt or DMA Ring Buffer interrupt). My ultimate goal is to write a driver that I can use for closed-loop control at 500 Hz using all 80 channels of the NI PCI-6255. I may also need to synchronize each scan with a NI PCIe-7841R card for which I've already written a driver. I want an interrupt-driven solution (be it programmed I/O on an interrupt or DMA that generates an interrupt) so that the CPU is available to other threads while the 80 analog inputs are being read (since it takes quite a while). I also want to minimize the number of interrupts. Basically, I will need to collect one sample from all 80 channels every 2 milliseconds.
    There are many different options available to do so, but what is the recommended technique for the NI PCI-6255 card? I tried using the AI FIFO interrupt without DMA, but it seems to interrupt as soon as any data is in the AI FIFO (i.e. not empty condition), rather than when all 80 channels are in the FIFO, so more interrupts are generated than necessary. I tried using DMA in Ring Buffer mode to collect a single sample of 80 channels and interrupting on the DMA Ring Buffer interrupt, which appears to work better except that this technique runs into problems if I cannot copy all the data out of the DMA buffer before the next AI scan begins (because the DMA will start overwriting the buffer as it is in ring buffer mode). If the DMA is not in ring buffer mode or I make the ring buffer larger than one 80-channel sample then I don't have a way to generate an interrupt when one sample has been acquired (which I need, because I'm doing control).
    I saw something in the documentation about a DMA Continue mode in which it looks like you can switch between two different buffers (by programming the Base Count/Address with a different address than the current address) automatically and thereby double-buffer the DMA but there is no real documentation or examples on this capability. However, I think it would work better than the Ring Buffer because I could interrupt on the DMA CONT flag presumably and be copying data out of one buffer while it is filling the other buffer.
    Another option would be DMA chaining, but again, I cannot find any information on these features specific to the NI DAQs.
    I tried interrupting on AI STOP figuring that I could get a single interrupt for each scan, but that doesn't appear to work as expected.
    I know that DAQmx on Windows has the ability to do such single sample, multiple channel tasks at a fixed rate so the hardware must support it.
    Any suggestions would be appreciated.
    Thanks.
    Daniel Madill

    Hello,
    The interrupt that will happen nearest the times that you need is the AI_Start_Interrupt in the Interrupt_A group. This interrupt will occur with each sample clock. By the second time this interrupt fires, the AI FIFO should have the samples from the first conversion. If it is easier to use programmed IO, you can read the samples out of the FIFO until you get all 80.
    Additionally, you can set the DMA to send samples as soon as the FIFO is no longer empty...instead of waiting for half full or full. This change will reduce latency for your control loop. You can set AI_FIFO_Mode in AI_Mode_3_Register to 0. By the second time this interrupt fires, you should be able to check how much data is in the DMA ring buffer and read the 80 samples when they are available. You can make the ring buffer larger than 80 samples if you see data getting overwritten.
    There is no interrupt associated with 80 samples being available in the FIFO or 80 samples being available/transferred by DMA to the host. X Series has much more flexibility with these interrupts.
    I hope this helps!
    Steven T.

  • What is the best approach to delete multiple list items

    Hi
    I want to delete multiple list items at once, what is best approach please suggest?
    Ramesh

    Hi Ramesh,
    Better approach for removing many items from a list is to use the ProcessBatchData method provided by the SPWeb class. First, you need a collection of the list items
    you want to remove, and then you can use the Get-SPListItem function to retrieve a list of items that match a given criteria. Then build a batch string with resultant items and call the ProcessBatchData method.
    Here is the sample code.
    function Get-SPList($webUrl, $lstUrl)
    $webObj = Get-SPWeb -identity $webUrl
    $lstObj = $webObj.GetList($lstUrl);
    return $lstObj
    $spList = Get-SPList -webUrl "http://mysitecollecton/MySite" -lstUrl "http://mysitecollecton/MySite/Lists/EmpInfo"
    $spQuery = new-object Microsoft.SharePoint.SPQuery
    $camlQuery =
    "<Where>
    <Eq>
    <FieldRef Name='Department' />
    <Value Type='Text'>HR</Value>
    </Eq>
    </Where>"
    $spQuery.Query = $camlQuery
    $spQuery.RowLimit = 100
    $spListItemCollection = $spList.GetItems($spQuery)
    # Create batch remove CAML query
    $batchRemove = '<?xml version="1.0" encoding="UTF-8"?><Batch>';
    # The command is used for each list item retrieved
    $command = '<Method><SetList Scope="Request">' +
    $spList.ID +'</SetList><SetVar Name="ID">{0}</SetVar>' +
    '<SetVar Name="Cmd">Delete</SetVar></Method>';
    foreach ($item in $spListItemCollection)
    $batchRemove += $command -f $item.Id;
    $batchRemove += "</Batch>";
    # Remove the list items using the batch command
    $spList.ParentWeb.ProcessBatchData($batchRemove) | Out-Null
    Please don't forget to mark it answered, if your problem resolved or helpful.

  • What's the best way to create multiple windows?

    I'm trying to write a program with Text fields and buttons on it which will open a completely different window with Text Fields and Buttons. The best thing I can find is JInternalFrame which creates a Frame inside of the window I'm working with. What do I need to use in order to make them separate?

    Exception occurred during event dispatching:
    java.lang.IllegalArgumentException: adding a window to a container
         at java.awt.Container.addImpl(Container.java:336)
         at java.awt.Container.add(Container.java:228)
         at beve.Frame5.actionPerformed(Frame5.java:181)
         at javax.swing.AbstractButton.fireActionPerformed(AbstractButton.java:1450)
         at javax.swing.AbstractButton$ForwardActionEvents.actionPerformed(AbstractButton.java:1504)
         at javax.swing.DefaultButtonModel.fireActionPerformed(DefaultButtonModel.java:378)
         at javax.swing.DefaultButtonModel.setPressed(DefaultButtonModel.java:250)
         at javax.swing.plaf.basic.BasicButtonListener.mouseReleased(BasicButtonListener.java:216)
         at java.awt.Component.processMouseEvent(Component.java:3715)
         at java.awt.Component.processEvent(Component.java:3544)
         at java.awt.Container.processEvent(Container.java:1164)
         at java.awt.Component.dispatchEventImpl(Component.java:2593)
         at java.awt.Container.dispatchEventImpl(Container.java:1213)
         at java.awt.Component.dispatchEvent(Component.java:2497)
         at java.awt.LightweightDispatcher.retargetMouseEvent(Container.java:2451)
         at java.awt.LightweightDispatcher.processMouseEvent(Container.java:2216)
         at java.awt.LightweightDispatcher.dispatchEvent(Container.java:2125)
         at java.awt.Container.dispatchEventImpl(Container.java:1200)
         at java.awt.Window.dispatchEventImpl(Window.java:914)
         at java.awt.Component.dispatchEvent(Component.java:2497)
         at java.awt.EventQueue.dispatchEvent(EventQueue.java:339)
         at java.awt.EventDispatchThread.pumpOneEventForHierarchy(EventDispatchThread.java:131)
         at java.awt.EventDispatchThread.pumpEventsForHierarchy(EventDispatchThread.java:98)
         at java.awt.EventDispatchThread.pumpEvents(EventDispatchThread.java:93)
         at java.awt.EventDispatchThread.run(EventDispatchThread.java:85)
    I need to take off now but I'll check back later tonight. Thanks for the help everyone :)

  • Upload BPC 5.1 budget info to R3 - what does SAP recommend?

    Hi folks
    We are implementing BPC 5.1.
    Previously we have used R3 transaction 7KEX to input budget data from Excel, but the business would like an automated upload from BPC.  We would prefer not to create z-versions to work around the problem.
    Does anyone have experience with automated budget uploads to R3, especially from OutlookSoft or BPC?
    Regards
    Russel

    Hi Russel,
    Unfortunately at the current time there is no automated method for retraction from BPC however this is a custom development. Right now the best way to do this is to use the method which you describe below.
    1) Export your data from BPC 5.1 using the Data Manager
    2) Import into R/3 using the Excel Budget data input.
    Cheers,
    Scott

  • What is the best approach to capture TBOM's for a SAP SRM system/functionality?

    Hello SCN Community,
    It would be much appreciated if somebody could share some information about the following....
    What is the best approach to create TBOM's for a SAP SRM system? The SRM functionality is basically consisting out of multiple ABAP Web Dynpro's that are connected as a process via a SAP Portal (as is understand it). The entrypint to the SRM functionality is via the SAP Portal.
    Do I first have to create a link to the Portal via an SAP Web Application link in SOLAR01 and then start recording? Will it record only the portal objects or also the ABAP Web Dynpro objects?
    Do I have to list all the separate ABAP Web Dynpro's in SOLAR01 and use those as a starting point?
    I am myself more familair with more classical SAP ABAP ECC systems and transactions.  I could hardly find any information on the use of BPCA and the required TBOM's in the area of SRM.... Any help would be much appreciated!
    Kind Regards,
    Guido Jacobs

    Hi Guido,
    today was a new blog released, maybe this helps:
    BPCA - Powerful Risk Eliminator
    Best Regards,
    Christoph

  • Icm/conn_timeout - SAP recommended value

    Hello All
    1.what is the SAP recommendations for icm/conn_timeout at client level
    2.why 400 timeout error not creating any dumps in SRM server.
    3.Are BUYERs allows to open more than one session like sourcing, redistibution workload,contract many - BUYERs open more than one time. what is SAP recommendations here on ITS sessions.
    4. What is good practice to work on only one ITS session(only sourcing) at a time or BUYER can open more than one web menus at a same time
    Muthu

    I can't comment on your first two questions. Below are my understandings regarding opening more than one sessions at a time.
    3.Are BUYERs allows to open more than one session like sourcing, redistibution workload,contract many - BUYERs open more than one time. what is SAP recommendations here on ITS sessions.
    4. What is good practice to work on only one ITS session(only sourcing) at a time or BUYER can open more than one web menus at a same time
    Before SRM 7.0, when SRM is based on ITS technology, the each transaction is launched in-frame, instead of being in a new window. That to me can be seen as a potential way to discourage users to launch more than one transaction at a time. The main reason could be to avoid mutually exclusive actions, such as one transaction locking the other. Or even hindering the business data consistency, e.g. when data update happens nearly simultenously.
    However since SRM 7, the standard approach has changed in this regard by allowing (or forcing) each process to be launched from ES Portal in a new window. Hence allowing the user to have multiple sessions open at the same time. This could be due to the real-estate/space limitation of Portal (since multiple applications could be sharing it, such as HR, SRM, etc).
    In general, my take (NOT SAP offical recommentation) is that users should be given options to work on different transactions at the same time, as long as they are fully aware of the effect and consequences.

  • System is creating multiple transaction cubes at the backend for one BPCApp

    Hi
    in one of the planning application, we have done a full optimization 8 times , out of that 5 times it came as error and 3 times it has successful.
    Now when we are sending the data through the Input Schedule it's shoiwng success and also the Data Manager package is also shoiwng success with the data showing witing back to the data base.
    But when the user is showing the report there is no data.
    So when we went to check the back end, we saw 7 transactional cubes with different technical name and 1 transactional cube with a _B sign associated with it, showing as a virtual info provider.
    Since the multi provider is not getting changed due to ful optimization we tried to figure out the very recent transactional cube by double clicking on the same and when we try to find the data with respect to data send we couldn't find any data.
    What can be the reasons of  creating multiple transactional cubes for a BPC planning application. And how to do we know which is the current cube for the respective BPC Planning application after the last Full Optimization.
    Regards
    Krish

    Hi Nilanjan
    Wish you a very happy new year.
    I need to get  it clarifed with the team whether they have aborted in between the Full Optimization process. Becuase i remember in one of the earlier engagement , we have aborted the modify application application and it cretaed multipple cubes at the back and subsequently SAP has suggested to delete the other cubes keeping the correct one.Can you elucidate the relevance of the _B Virtual Info provider from the rationale perspective, why its required them.
    I shall update you on the same.
    Thanks
    Krish

  • Best Approach to create Security / Authorization Schema for an APEX Apps

    Hi,
    I am planning to create a Security / Authorization Schema for an APEX Application.
    Just want to know what is the best approach to create the security feature in APEX, so that it should be re-used in other APEXApplications too..
    I am looking for following features...
    1. users LOGIN and then user's name is stored in APEX_USER...
    2. Based on the user, I want to restrict the Application on following levels.
    - TABS
    - TABS - Page1 (Report
    - Page2 (Form)
    - Page2 (Region1)
    - Page2 (Region1, Button1)
    - Page2 (Region1, Items,....)
    AND so on.....basically depending on user....he will have access to certain TABS, Pages, Regions, Buttons, Items...
    I know, we have to create the Authorization Schema for this and then attach these Authorization Schema to the different Level we want.
    My Question is, what should be the TABLE structure to capture these info for each user...where we will say...this USER will have following access...AND then we create Authorization Schema from this table...
    Also what should be the FRONT end, we should have to enter these detail...
    SO, wondering, lot of people may already have implemented this feature....so if guys can provide the BEST Approach (re-usable for other APEX Application)....that will be really nice..
    Thanks,
    Deepak

    Hi Raghu,
    thanks for the detial info.
    so that means..I should have 2 table...
    master table (2 columns - username, password)
            username    password
       user1       xxxx
       user2       xxxx2nd table (2 columns - username, chq_disp_option)
    - In this table, we don't have Y/N Flag you mentioned..
    - If we have to enter all the regions/tabs/pages in the Applications here or just those regions/tabs/pages for which are conditionally diaplayed.
    - so that means in all the Pages/Regions/tabs/items in the entire Application, we have to call the Conditionally display..
    - suppose we have 3 tabs, 5 pages, 6 regions, 15 items..that means in this table we have to enter (3+5+6+15) = 29 records for each individual users..
              username    chq_disp_option
       user1       re_region1
       user1       re_region2
       user1       tb_main
       user1       Page1
       user1       Page5
       ----        ----     - how you are defining unique name for Regions..i mean in static ID or the Title
    - is the unique name for tab & item is same as the TAB_NAME (T_HOME) & Item Name (P1_ITEM1) or you are defining somewhere else.
    Thanks,
    Deepak

  • Best approach to creating a TOC for product catalog using data merge

    What is the best approach for creating a TOC for a product catalog (over 1,000 items) using Data Merge?
    The TOC would contain the product Categories. 
    So for example, Category A items could go from pages 1 - 3, and Category B items would start at pg 4, but if new items were added to Category A, then Category B may start from pg 6. 
    From the Data Source, there are 5 Data Fields I've chosen to be displayed.  If this were a regular digital print document, I could use the Paragraph Style method for creating a TOC, but if I make any one of the Data Fields a certain Paragraph Style and use that for the TOC, it'll populate the TOC with that Data Field for all the items. 
    Any suggestions?

    Peter Spier wrote:
    TOC is not interactive in the ID file, though it can be in a PDF that you export (there's a checkbox to create PDF bookmarks). You might want to think about using Cross-references (rather than hyperlinks, I think) to build the TOC. You have to do it manually, but once done it should maintain itself, whereas a TOC is built automatically, but must be regenerated after you edit the doc.
    One caveat witih TOCs created from cross-references: Although changing the text of an x-ref source paragraph (for example from "Patatas and tamatas" to "Tomatoes and Potatoes,"and/or when the source paragraph flows to the next or previous page) update automatically or when invoking "Update cross-references," MOVING a cross-reference source paragraph to a location before or after another source paragraph, does not change their sequence in the pseudo-TOC. You'll need to manually move the reference in the pseudo-TOC to the correct position in the sequence of cross-refs. So, put the task of checking the order of x-refs in the pseudo-TOC on your before hand-off check list.
    HTH
    Regards,
    Peter
    Peter Gold
    KnowHow ProServices

  • Creating Multiple Playback Positions on a Track

    Hello,
    This should be quite simple and i'm trying to emulate a feature on youtube. Basically, i have some very long music tracks, mixes and so forth, that are about an hour long. What is the simplest way to create multiple playback positions in one track? This is sort of like how you can list multiple playback positions on youtube. I'm not sure if this is the same as remembering playback positions, as that is more about playing from when you left off. I'm talking more clickable playback positions for a single music track/mix.
    Thanks,

    I'm not an expert but I think that is only possible with video files which support chapter markers.  I have never heard of anybody doing it in iTunes or with music files.  The closest I have heard is iTunes allowing you to set a start and stop time for tracks, but I think that is saved in iTunes data and is not part of the file.

  • Creating extra send buttons on the channel strips

    I am having a difficult time in finding a way to add mor sends than the two (2) sends I currently have. I need more, at least four (4). I believe that there are eight (8) available in Logic Pro 7.1. What is the procedure to add (create) additional sends. I have checked the manuals, and can find nothing pertaining to this. I could use some help here.
    Thanks - JJFournet

    Like, "duh", I figured it out last night! Thanks for your response, though.
    And thanks for the responses from the other helpful people.
    JJFournet

  • How to create multiple files with Receiver File Adapter in SAP PI 7.31 Java Stack

    Dear Friends,
    I am using Sender JDBC Adapter and Receiver File Adapter in Integration Flow in SAP PI 7.3 EHP 1 SP08 Java Stack environment. The requirement is that we need to create multiple files based on the row count in jdbc resultset. If there are 5 rows in resultset, we need to create 5 XML files with one row elements in one file. Similarly if there are 10 rows, we need to create 10 XML files.
    So how can we create multiple files in this scenario. I tried placing a for loop in the Java Mapping as below in the transform method:
    DynamicConfiguration conf = arg0.getDynamicConfiguration();
    StringBuffer sbFileData = new StringBuffer();
    for (int i =0; i < record.size(); i++ {
         . // Create XML for each row and Marshal the object into to the String Buffer
         String strFileName = "DC_" + new SimpleDateFormat("ddMMyyyyHHmm").format(new java.util.Date())+"_"+i+".xml";
         conf.put(KEY_FILENAME, strFileName);
         arg1.getOutputPayload().getOutputStream().write(sbFileData.toString().getBytes("UTF-8"));
         arg1.getOutputPayload().getOutputStream().flush();
    So here I'm flushing the OutputStream for each record. But it's not creating the multiple file, instead it creates only one file will all record XMLs appended to each other.
    Please let me know if I missing something or need to do some thing else.
    Regards,
    Shreyansh Shah

    Hi
    You can easily achieve this using graphical mapping.  Create your target message type like below
    MT_Target
      Details  0 to 1
          Data  0 to 1
    Source sample structure
    <resultset>
    <row>
    <column-name>column-value</ column-name>
    </row>
    Then do the message mapping like below
    map <row> with  MT_Target
    contant ----> Deatils
    column-name ------>Data
    In the signature tab of message mapping, choose the occurrence of your target message type as
    0 to unbounded.
    This will generate multiple files from multiple rows.
    Let me know if you have any doubt.

Maybe you are looking for

  • WIndows 8.1 app tiles don't work since window update on 12/14/14

    My computer with windows 8.1 went through several updates on 12/14/14 and now none of my window tiles will open and several are now missing. I tried to install them again, and it says go back to the app store and install them again. It doesn't work.

  • Backing up virtual machines in Hyper-V: theory question

    Hello! This article http://msdn.microsoft.com/en-us/library/dd405549%28v=vs.85%29.aspx says: "Hyper-V uses the Volume Shadow Copy Service (VSS) to backup and restore virtual machines (VMs). If the backup (volume snapshot) integration services are ins

  • Can I buy a new power cable?

    My MacBook power cable is falling appart; the plastic is wearing away and it will only charge when it's positioned in certain ways. I was wondering if there was anyway I could buy a new power cable for my MacBook and how much it would cost. I would r

  • Vendor login

    I have tried logining in to vendor portal, but it says account is disabled. There is no contact information, or information of any kind to find out why and what to do. I have tried creating a new account, but it says company name already exists.  Is

  • My Mac is waking up from sleep without reason.

    So this is what happens: Sometimes I send my iMac to sleep and then I go away, but after I return, I've noticed the hard drive is working as if the computer were awake, not sleeping. For some reason it is waking up. Sometimes I'm sitting at my office