Set relation between list and document library?

I have a requirement that I want to show pop up/ modal dialog when I click on list item column "upload/Edit" (in EditForm.aspx) . In Modal Dialog I have to show upload button and also existed items if any. Here the attachments will store into document
library's folder .The list item ID should be the folder name in document library.
How do I set relation between list and doc lib ?
Appreciate if suggests any links.
Thank you.

Hi Saurav,
pls check below
What version of SharePoint 2010 "SP 2010 Foundation or SP 2010 Server" you need to have SP 2010 server and to have the "User Profile service application".
How do you configure the "User Profile service application" did you add the "Social Tagging Database"
Create, edit, or delete a User Profile service application (SharePoint Server 2010)" 
http://technet.microsoft.com/en-us/library/ee721052.aspx
When you access to "Application Management>Manage service applications>User Profile Service Application>Manage User Permissions" validate if the all authenticated users Group have the "use Social Features" checked.
Also validate in Central Administration in Farm Features if you have social tags and notes activated.
http://sharepoint.stackexchange.com/questions/17546/tagging-feature-not-working
https://social.technet.microsoft.com/Forums/office/en-US/c11cda96-091b-4b96-91bc-ccd8000238f4/tags-and-notes-sharepoint-2010-not-visible?forum=sharepointadminprevious
Please remember to click 'Mark as Answer' on the answer if it helps you

Similar Messages

  • Sharepoint 2013 on premises Tags & Notes button in List and Document library is disabled.

    Hi,
    In My Sharepoint  2013 on premises  installation Tags & Notes button in List and Document library ribbon is appearing as greyed out.
    I have checked that managed metadata service, User profile services are running. Also have given required permissions to the logged in user.
    As I came to know that Tags & Notes feature has been retired in Sharepoint online. Is this happening because of same reason as I have downloaded the Sharepoint 2013 on premises version recently or do I need to change some settings.
    Please advice.

    Hi Saurav,
    pls check below
    What version of SharePoint 2010 "SP 2010 Foundation or SP 2010 Server" you need to have SP 2010 server and to have the "User Profile service application".
    How do you configure the "User Profile service application" did you add the "Social Tagging Database"
    Create, edit, or delete a User Profile service application (SharePoint Server 2010)" 
    http://technet.microsoft.com/en-us/library/ee721052.aspx
    When you access to "Application Management>Manage service applications>User Profile Service Application>Manage User Permissions" validate if the all authenticated users Group have the "use Social Features" checked.
    Also validate in Central Administration in Farm Features if you have social tags and notes activated.
    http://sharepoint.stackexchange.com/questions/17546/tagging-feature-not-working
    https://social.technet.microsoft.com/Forums/office/en-US/c11cda96-091b-4b96-91bc-ccd8000238f4/tags-and-notes-sharepoint-2010-not-visible?forum=sharepointadminprevious
    Please remember to click 'Mark as Answer' on the answer if it helps you

  • Configure incoming mails to lists and document library

    Hi Everyone,
    I am trying to configure incoming mail in Sharepoint to accommodate sending emails to a list or a document library. Between the basic/automatic mode and advanced mode, I have opted to configure basic as I do not want to yet have to configure any settings
    within our Exchange environment. I have set the incoming email server display address to be Mail.Sharepoint.com and specified that all emails should be accepted. I have also gone on to our Sharepoint web server and enabled SMTP service. I apart from enabling
    the smtp service I have not configured anything else on the web server. Finally I have gone into the incoming email settings of a test document library and assigned  the name [email protected]
    Having tried multiple times sending mails to the above mentioned address and getting the error "#,#5.7.0 smtp; 551 5.7.0 recipient domain does not exist in DNS>#SMTP#" please could you suggest what the problems is?
    Thanks,
    Dominic

    You need to add an A record in DNS of mail.sharepoint.com to the DNS Zone hosting 'sharepoint.com'. You could also consider adding an MX record that points to that A record.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Table relations between vbrk and bkpf for  Accounting Document Number

    hello,
    i am using 4 tables to get data into my programs.
    vbrk,vbrp konv and bkpf.
    i want to get belnr from bkpf.i found relation between vbrk and belnr.but in vbrk table belnr's value is initial.
    can anybody tell me that how should i relate vbrk and bkpf or how to get Accounting Document Number(belnr) from bkpf for Billing Document(vbeln).
    regards,
    soniya s.

    hi,
    chekc this. its working for me.
    data : WA_AWKEY LIKE BKPF-AWKEY.
    data :  WA_BELNR LIKE BKPF-BELNR.
    data : LENGTH TYPE I.
    *BREAK MTABAP.
    LENGTH = STRLEN( IT_VBRK-VBELN ).
    if  LENGTH = '10' .
    MOVE it_vbrk-VBELN TO WA_AWKEY.
    SELECT SINGLE BELNR FROM BKPF INTO WA_BELNR
      WHERE AWKEY = WA_AWKEY
      AND AWTYP = 'VBRK'
      and blart = 'RV'.
    it_final-acc_doc = WA_BELNR.
      CLEAR WA_BELNR .
      CLEAR WA_AWKEY .
    else.
    CONCATENATE '0' it_vbrk-vbeln INTO wa_awkey.
    SELECT SINGLE BELNR FROM BKPF INTO WA_BELNR
      WHERE AWKEY = WA_AWKEY
      AND AWTYP = 'VBRK'
      and blart = 'RV'.
    it_final-acc_doc = WA_BELNR .
      CLEAR WA_BELNR .
      CLEAR WA_AWKEY.
    endif.

  • Relation between BKPF and EKBE table for reversing document?

    Is there any relation between BKPF and EKBE table for reversing document when STBLG is blank in BKPF.
    AND
    How to differentiate Old Accounting document / Reverse document / New document when in BKPF-STBLG is blank for all three document and BKPF-XBLNR are same for all.

    Hi Varun,
    Table Document no (BKPF-BELNR) is link between Material document no (EKBE-BELNR). If you are looking for validating Material reversal document then please use following logic for your requirement.
    Material Document (MBLNR) & Material Document Year (MJAHR). Pass material document (MBLNR) to MSEG-SMBLN to read MSEG-MBLNR with MSEG-MJAHR to MSEG-SJAHR, if record found delete both the documents i.e. MSEG-SMBLN & MSEG-MBLNR (donu2019t consider these documents in the logic) and display Material Document No. (MSEG-MBLNR)
    Regards,
    Santosh

  • SharePoint 2013 users only see custom list and document lists that they uniquely created

    In my first SharePoint 2013 project, I want to create a view where users can only see their own files that they generated. I think I know how to create this view when the list is associated with the 'Documents; folder. However I have created several custom
    lists where I can only see these items when I am on the navigation pane to the right and the unique custom lists are listed under 'recent'.
    Let me know if you can answer any of the following questions:
    1. When on the main SharePoint website, can lists be displayed in other locations than what I just listed? If so, where would I see these other lists?
    2. I would like to include the custom lists that are located in the 'recent' navigation pane as a view that only I can use. This would be based upon userid. If this is possible,  how would you combine the lists lists in the same view? If this is not possible,
    how would you set a view just for the lists listed under the 'recent' section?
    3. The content approval solution under versioning settings of a library to restrict the documents to only ones that the user created (and also Approver rights).  Then if they create a new view they still won't see the documents that were not created by
    them. What about custom lists that are not in the documents area but are in the 'recent' area? Would this restrict users to see only items that they created?
    4. Item level Permissions in advance settings solve this problem for the list in the document area and the custom list in the recent area? If so, how would you setup these values?

    As it turns out the, setting Item-Level Permissions in a library is fully supported with PowerShell!
    The PowerShell commands for changing this are very simple:
    $web = Get-SPWeb http://YourSite/
    $list = $web.Lists[“Your Document Library Name”]
    $list.ReadSecurity = 2
    $list.Update()
    $web.Dispose()
    Note the 3rd line which is where you determine the value for this setting using the following values:
    1 = “Read all items”
    2 = “Read items that were created by the user”
    If you wish to modify the values for Create and Edit access instead, replace .ReadSecurity with .WriteSecurity with
    the following values:
    1 = “Create and edit All items”
    2 = “Create items and edit items that were created by the user”
    4 = “None”
    For example:
    $web = Get-SPWeb http://YourSite/
    $list = $web.Lists[“Your Document Library Name”]
    $list.WriteSecurity = 2
    $list.Update()
    $web.Dispose()
    Please 'propose as answer' if it helped you, also 'vote helpful' if you like this reply.

  • Relation between idoc and EDI,ALE

    Hi,
           I am new to IDOC.  Can u please tell me whether there is any relation between IDOC and ALE or IDOC and EDI. If so what is the relation between them.

    Hi Buvana,
    ALE is a means of migrating DATA between SAP systems. You configure logical destinations, establish some migrating criteria and the system takes care of getting your data from point A to point B.
    EDI ->
    EDI is described as the interchange of structured data according to agreed message standards between computer systems, by electronic means. Structured data equates to a simple and direct method of presenting the data content of a document. The method of ensuring the correct interpretation of the information by the computer system is defined by the EDI standard."
    EDI is a technique used to communicate business transactions between computer systems of different companies and organizations. Note that sometimes the EDI mechanism deployed at a company is often used to interface to other systems within the same organization."
    Info shuttle - "I was implementing EDI at a customer and I wanted some real data to test with in Development, so I asked the Project Manager to see if he could arrange something. Literally, 20 minutes later he came back and said that there were 10 new orders in Development, together with the customer master records related to the order, the carrier vendor master record, the material master records, ... were all there. I asked him how he had done it so quickly which led me to my first exposure with Infoshuttle, a tool that does this for you using the ALE functionality. As an ALE consultant I know how to do this stuff but I also know how long it would take to set that up.
    www.sapgenie.com/sapedi/edi_sap_training.htm
    www.sap-img.com/basis/ difference-between-edi-and-idoc.htm
    www.help.sap.com/saphelp_nw04/helpdata/ en/35/26b592afab52b9e10000009b38f974/content.htm
    www.help.sap.com/saphelp_nw04/helpdata/ en/35/26b594afab52b9e10000009b38f974/content.htm
    http://www.onestopsap.com/interview-Question/edi/
    IDocs act as data containers in an ALE scenario. They are just a format of storing application data and then transferred using ALE. Take a look at http://www.sapgenie.com/ale/why_ale.htm and http://www.sapgenie.com/sapgenie/docs/ale_whitepaper.doc for further details on ALE.
    Check this link.
    http://www.sapgenie.com/sapedi/idoc_abap.htm
    http://www.henrikfrank.dk/abapexamples/IDOC/IDOC.htm
    http://help.sap.com/printdocu/core/Print46c/en/Data/Index_en.htm
    http://help.sap.com/saphelp_47x200/helpdata/en/0b/2a6095507d11d18ee90000e8366fc2/frameset.htm
    Regards,
    Priyanka.

  • How to get data from three tables (A,B,C) having one to many relation between A and B .and having one to many reation between b and c

    i have  three tables A,B,C.  there is one to many relation between A and B. and one to many relation existed between table b and c . how will get data from these three tables

    check if this helps:
    select * --you can always frame your column set
    from tableA a
    left join tableB b on a.aid=b.aid
    left join tableC c on c.bid=b.bid
    This is just a general query. However, we can help you a lot more, if you can post the DDL + sample data and required output.
    Thanks,
    Jay
    <If the post was helpful mark as 'Helpful' and if the post answered your query, mark as 'Answered'>

  • What are the Relations between Journalizing and IKM?

    What is the best method to use in the following scenario:
    I have about 20 source tables with large amount of data.
    I need to create interfaces that join the source tables into target tables.
    The source tables are inserted every few secondes with about hundreds to thousands rows.
    There can be a gap of few seconds between the insert of different tables that sould be joined.
    The source and target tables are on the same Oracle instance and schema.
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?
    In general What are the relations between 'Journalizing' and 'IKM'?
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?
    I want to understand what is the role of 'Journalizing CDC'?
    Can 'IKM - Incremental Update' work without 'Journalizing'?
    Does 'Journalizing' need to have PK on the tables?
    What should i do if i can't put PK (there can be multiple identical rows)?
    Thanks in advance Yael

    Hi Yael,
    I will try and answer as many of your points as I can in one post :-)
    Journalizing is way of tracking only changed data in your source system, if your source tables had a date_modified you could always use this as a filter when scanning for changes rather than CDC, Log based CDC (Asynchronous in ODI, Logminer/Streams or Goldengate for example) removes the overhead of of placing a trigger on the source table to track changes but be aware that it doesnt fully remove the need to scan the source tables, in answer to you question about Primary keys, Oracle CDC with ODI will create an unconditional log group on the columns that you have defined in ODI as your PK, the PK columns are tracked by the database and presented in a Journal table (J$<source_table_name>) this Journal table is joined back to source table via a journalizing view (JV$<source_table_name>) to get the rest of the row (ie none PK columns) - So be aware that when ODI comes around to get all data in the Journalizing view (ie Inserts, Updates and Deletes) the source database performs a join back to the source table. You can negate this by specifying ALL source table columns in your PK in ODI - This forces all columns into the unconditional log group, the journal table etc. - You will need to tweak the JKM to then change the syntax sent to the database when starting the journal - I have done this in the past, using a flexfield in the datastore to toggle 'Full Column' / 'Primary Key Cols' to go into the JKM set up (there are a few Ebusiness suite tables with no primary key so we had to do this) - The only problem with this approach is that with no PK , you need to make sure you only get the 'last' update and in the right order to apply to your target tables, without so , you might process the update before the insert for example, and be out of sync.
    So JKM's provide a mechanism for 'Change data only' to be provided to ODI, if you want to handle deletes in your source table CDC is usefull (otherwise you dont capture the delete with a normal LKM / IKM set up)
    IKM Incremental update can be used with or without JKM's, its for integrating data into your target table, typically it will do a NOT EXISTS or a Minus when loading the integration table (I$<target_table_name>) to ensure you only get 'Changed' rows on the load into the target.
    user604062 wrote:
    I want to understand the role of: 'Journalizing CDC' and 'IKM - Incremental Update' and
    how can i use it in my scenario?Hopefully I have explained it above, its the type of thing you really need to play around with, and throroughly review the operator logs to see what is actually going on (I think this is a very good guide to setting it up : http://soainfrastructure.blogspot.ie/2009/02/setting-up-oracle-data-integrator-odi.html)
    In general What are the relations between 'Journalizing' and 'IKM'?JKM simply presents (only) changed data to ODI, it removes the need for you to decide 'how' to get the updates and removes the need for costly scans on the source table (full source to target table comparisons, scanning for updates based on last update date etc)
    Should i use both of them? Or maybe it is better to deelte and insert to the target tables?Delete and insert into target is fine , but ask yourself how do you identify which rows to process, inserts and updates are generally OK , to spot a delete you need to compare the table in full, target table minus source table = deleted rows , do you want to copy the whole source table every time to perform this ? Are they in the same database ?
    I want to understand what is the role of 'Journalizing CDC'?Its the ODI mechanism for configuring, starting, stopping the change data capture process in the source systems , there are different KM's for seperate technologies and a few to choose for Oracle (Triggers (Synchronous), Streams / Logminer (Asynchronous), Goldengate etc)
    Can 'IKM - Incremental Update' work without 'Journalizing'?Yes of course, Without CDC your process would look something like :
    Source target ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    With CDC your process looks like :
    Source Journal (J$ table with JV$ view) ----< LKM >---- Collection table (C$) ----<IKM>---- Integration table (I$) -----< IKM >---- Target table
    as you can see its the same process after the source table (there is an option in the interface to enable the J$ source , the IKM step changes with CDC as you can use 'Synchronise Journal Deletes'
    Does 'Journalizing' need to have PK on the tables?Yes - at least a logical PK in the datastore, see my reply at the top for reasons why (Log Groups, joining back the J$ table to the source table etc)
    What should i do if i can't put PK (there can be multiple identical rows)? Either talk to the source system people about adding one, or be prepared to change the JKM (and maybe LKM, IKM's) , you can try putting all columns in the PK in ODI. Ask yourself this , if you have 10 identical rows in your source and target tables, and one row gets updated - how can you identify which row in the target table to update ?
    >
    Thanks in advance YaelA lot to take in, as I advised I would reccomend you get a little test area set up and also read the Oracle database documentation on CDC as it covers a lot of the theory that ODI is simply implementing.
    Hope this helps!
    Alastair

  • Relation between URL and content in JSF navigation

    Hello,
    I've tried lots of JSF examples and there always occoured a strange kind of navigation to me:
    The URL you see in the browser's address bar has nothing to do with the actually displayed JSP. Instead of that, the URL is always pointing to the previously called JSP - where the call to the current JSP comes from.
    I know where this comes from:
    If the form tags don't define an action, they are pointing to the current page itself.
    And the actually called navigation case is chosen by a (hidden) post parameter.
    So the JSF controller servlet doesn't actually care, which URL is called - it just regards the mentioned parameter.
    This kind of navigation seems quite confusing, especcially since JSF claims to make navigation clearer.
    One of the basic demands of the web applications we are developing is, to have a clear relation between URL and page content.
    E.g. it must be possible to index the pages with search engines. Or to simply bookmark distinct pages in the browsers favourites.
    Another problem about this navigation concept is, that it always works with forms and with post method (you can't choose get method), even if you just have simple links - e.g. detail view of a single item after selecting it from a list of items. That way it's not possible to ever use the back button of your browser.
    I'm new to JSF, so my question:
    Is this the (one-and-only) navigation concept of JSF, or are there alternative concepts as well?
    I would be surprised, since the mentioned demands are basic web standard - although I didn't find any alternatives so far.
    Thanks,
    Chily

    Thanks for the answer!
    I'm not sure, if using a redirect on every single call is very achieveable and in the purpose of the JSF design.
    E.g. it is generating a lot of traffic overhead.
    (Identifying the navigation case by the page name instead of a hidden parameter would help a lot.)
    Or the history of the browser doesn't work with it, so you can't identify visited links.
    But anyways, there are also internal problems with it in most cases: I've inserted the <redirect/> tag in the navigation cases of a few examples. This just causes a call of the <to-view-id> page, without any parameters. So I probably have to add some non-standard-processing around the call to make it work:
    If e.g. I have the standard case of a list of items (let's say in a dataTable), which are identified by an id. And each item uses a commandLink to call it's detail view.
    * How do I get the identifier (id) as a parameter into the redirect URL?
    * And how do I access this parameter in the destination form bean to choose the item to be displayed? (see also http://forum.java.sun.com/thread.jspa?threadID=607505)
    Thanks for help,
    Chily

  • Relation between GR and Theoritical qty

    Hi,
    I have an issue with a report program which fetches the PO and material data from ME23n.The program should display PO, material head, material quantity, GR qty, vendor, theoretical qty, Actual qty.
    Here for one of the POs, the GR qty is 1CU but theoretical qty is getting as 8640.
    Can anyone explain what is the relation between GR and theoretical qty ? and how to solve this issue?
    How to calculate theoretical qty ?
    Thanks.

    Hi
    By GR you mean Goods Receipt.There is no Idoc for this process.SRM uses XML format for Confirmation which is equivalent to Goods confirmation process in ERP system.
    GR Idoc - You can try WE09. Give the message type and all the details that you have. you will get the Idoc number.
    If you are in position to tell what is your exact requirement with table in SRM system,i should be in position to help.
    There is a transaction called BBP_PD using which you can get the required table.
    Just go to this transaction in SRM EASY Access Menu.
    Select Object Type as (Take F4 Help here)
    for confirmation ,object type = BUS2203.
    You will get an interactive list and you can browse the list and in the last node ,you will get some table names which may be helpful.
    If you have the PO# you can use the FM BBP_PD_CONF_GETLIST with the parameter I_FOR_PO_ID, this will return a E_PDLIST table with the GR numbers related to the PO that you used as input.
    Useful related pointers ->
    Goods Receipt and Invoice Receipt
    Re: interface required to post invoice in EBP system in SRM4.0
    Re: BAPI or FM to create Invoice in SRM , without reference to PO ?
    Do let me know.
    Regards
    - Atul

  • Relation between ODPI and OINV

    Hello all,
    I have a problem. I want to make a report in Crystal report. This report will show a list of invoice and downpayment which is related to this invoice. I dont know how to get downpayment which is use on invoice. I dont find relation between invoice and downpayment.
    Can you help me?

    Hi Vit,
    I used SQL Profiler to extract the query that is used to get the records when the ... button is pressed. I am not sure if this is what you're looking for, but here is the query:
    SELECT T0.* 
    FROM [dbo].[ODPI] T0
    WHERE (T0.CardCode = N'CARDCODE       ' 
    OR  T0.FatherCard = N'CARDCODE       ' 
    AND  T0.FatherType = N'P' )
    AND  T0.CANCELED = N'N' 
    AND  T0.DpmStatus = N'O' 
    AND  T0.DocCur = N'GBP' 
    AND  T0.PaidSum <> 0 
    AND  T0.DocStatus = N'C' 
    AND  T0.Posted = N'Y'
    Hope it helps,
    Adele

  • Relation Between RSEG and BSET Tables

    Hi ,
      I want to Calculate the the tax for MIRO document Line item wise, i am not able to get the relation between RSEG and BSET from where i can get the taxes. so, any one can help me to resolve this issue on priority.
    Regards,
    Venugopal.

    Hi, Mr. fantasy,
    I'm not too familar with MIGO - now working in non-logistics environment.
    Did you already try RSEG fields
    LFBNR     LFBNR     CHAR     10     0     Document No. of a Reference Document                                  
    LFGJA     LFGJA     NUMC     4     0     Fiscal Year of Current Period                                  
    LFPOS     LFPOS     NUMC     4     0     Item of a Reference Document     
    just an idea.
    Regards,
    Clemens

  • Relation between master and transaction data

    Hi experts,
    I have a basic question how the relation between masterdata and transction data is maintained in R/3. As I know first we extract master data and later transaction data in a typical R/3 based scenario.
    1.How does BW knows that , the transaction data are  related masterdata in the same sense as in R/3 ?
    I am involved in an enhancement of R/3 extractor. I need to derive 3 more fields and map it to info objects in BW. Cld any one explain me ( with an example plz) how these alignments to be made. How can I establish relation b/w derived fields and transaction data during extractions ?
    regs
    D Bret

    Hi Dave.
    Let's see if I can get all of your questions answered.  First, when Transactional Data is loaded, BW checks to see if a SID exists for each characterisatic field & value in the load.  If one doesn't, it either creates a SID (Surregate ID) in the master data tables (default action) or the load fails (if the enforece referential intgegrity flag is set).  It's better to load master data first in either case as it will improve the performance of the transactional load.
    In terms of R/3, the data has similar relationships as in BW.  There are master data tables and transactional data tables.  The difference is that while R/3 uses ~1 master data table per field, BW uses more.  Same thing for transactional data.
    As for your enhancement, if everything you need for the derivation of the extra field is in R/3, then you can enhance the extractor in RSA6 and do the mapping in R/3.  Once you replicate the DataSource to BW and reconnect the transformation & update rules, the load should work.  If you can give a bit more information on what sort of enhancement you are attempting, I can give a better example.
    Cheers,
    Adam

  • Relation between GDM and dtlogin ?

    Hi !
    I'm wondering what is the exact relation between gdm and dt ? I assume that on a fresh (vanilla) install of Solaris 10 (x86 and Sparc, as I've tried both) the only reference to a graphical login is /etc/init.d/dtlogin. How then is gdm invoked ? The login invite seems to be a gdm one.
    I've tried modifying gdm.conf (manually and with gdmsetup) without seeing any modifs taking effect, even after a reboot the following day.
    What am I missing badly ? :)
    On a side-note, I'm trying to setup XDMCP so that other boxen will be able to graphically login onto that Sun box.
    Thanks,
    Jeff

    Dear Phani,
    http://help.sap.com/saphelp_erp2005/helpdata/en/70/a7853478616434e10000009b38f83b/frameset.htm
    VBUK .... Sales Document: Header Status and Administrative Data
    VBUP ....Sales Document: Item Status
    VBUK contains header status where as VBUP contains individual item status of the header. Similar to VBAK and ABAP where VBAK contains Sales Document header data and VBAP contains Sales document item data.
    Hope this will help.
    Regards,
    Naveen.

Maybe you are looking for

  • How to add a java library?

    Hello! Can some one help me on the following problem? I am writing a DOM parser and I need to print out the modified xml file. So I want to use the write funcition defined in the com.sun.xml.tree.XmlDocument library. But this package is NOT included

  • How do you make drop caps in Pages?

    Word has a drop down in the Format column that give you an ap to make drop caps... like at the beginning of a chapter. How do you do it in Pages?

  • HGTV in widescreen only

    Noticed today that HGTV is in widescreen only mode in hd , and sd.  The wife hates it because of the black boxes above and below.  We have tried to correct this in the box settings and the tv settings.  Anybody else notice this?  We are in southern C

  • Output screen basic list buttons

    Hi frnds,                what are the steps to create buttons in the application toolbar in our output screen. like i m getting sales header data in my basic list so in the output i m having a button which i press then i will get sales item details.

  • Keyboard Turns On and Off every 40-50 seconds

    I'm using the aluminium macbook now.. the one with the most basic specs. Lately as I play Warcraft (not World of Warcraft) I find that my keyboard does not respond after about 1 minute or so into the game... then it works for a few seconds and it doe