Tabular Model question

Hi,
I'm starting to use SSAS tabular mode and have a question.
I created a model and named it Subscriptions and was now making a report to query it. However it was deployed as Model. I went to the project properties and noticed that the Cube Name was set to Model.
This means that I can only have 1 model per SSAS Tabular project? I tried to add a new Tabular model and it says I can only have 1 bism file in a project.
Thanks

It's a balance. Fact tables or subject areas that will be queried together or side by side should be in the same model. Unrelated subject areas can go in a different database if you like. Yes, you are balancing duplicating dimensions across several models
with the complexity of managing one huge model.
http://artisconsulting.com/Blogs/GregGalloway

Similar Messages

  • DAX Tabular model - Question regarding filtering twice

    We have a Fact table with two data sources. Now I have a DAX measure to filter out to one data source
    Chat Handled := CALCULATE( SUM([Handled]),DimDataSource[SourceOrType] = "Chat Source 1")
    I have other Fact tables where we have other data sources - which are defined in the DimDataSource. So basically when we try to conform across all Facts with DataSources, other Facts are looking good - expect this Chat Fact where it shows the same numbers
    again.  Below is the view when I slice the data using a pivot -
    Is there a way to fix this (mainly for chat handled) above  or is there any work around for this ? Appreciate your help with this.
    Row Labels
    Incidents
    Chat Handled
    Surveys
    Phone calls
    Incidents  Source 1
    123,456
    12,345,678
    Survey  Source 1
    12,345,678
    123,456
    Incidents  Source 2
    12,345,678
    12,345,678
    Survey  Source 2
    12,345,678
    234,567
    Survey  Source 3
    12,345,678
    345,678
    Survey  Source 4
    12,345,678
    456,789
    Phone  Source 1
    12,345,678
    1,234,576
    Phone  Source 1
    12,345,678
    2,345,678
    Chat  Source 1
    12,345,678
    Chat  Source 2
    12,345,678

    The only thing I can think of with the information you gave is that there is something wrong with the relationship between the Fact and your Dim.
    Is the relationship defined in the data model ? (it seems to me that filtering the Dim does not propagate to the Fact)
    Driezl

  • Question on Tabular Model 2012

    Hi,
    I need to hide certain columns for a user group but the same columns should be visible for another user group in the same model. The out of the box 'row level  security' would not work for this requirement. I tried a modelling technique, by bringing
    the 'hidden fields' to a table using the 'Blank' DAX function in a 'role' on the table that contains the hidden fields, but this hides all tables related to the table that contains the hidden fields.
    As a last resort I am creating two models (1) model one with all fields visible (2)  model two will hide certain fields. Basically model 2 is a copy of model 1, only certain fields in model 2 will be hidden.
    To avoid reading the source data twice I am planning to update the model 1 and then synchronize model 1 with model 2 ( the one with hidden fields).  My gut feeling is that if I synchronize model 1 with model 2, all the hidden fields in model 2
    will become visible, as the metadata in model 2 will be updated with metadata of model 1.
    As the number of users would be in hundreds, implementing a complex security layer might impact the performance if we try and implement such security on the same model.
    It would be great if anybody can share thoughts on the probable solution and if the approach to 'synchonize' both the models would work.
    Thanks
    Bidyut
    A**LE

    Hi Bidyut,
    According to your description, you want to implement dynamic security for two group users, right? In SQL Server Analysis Services Tabular model, we can implement dynamic security using an external security table but we can integrate this security table into
    the existing tables based on particular source data at production implementations. So in your scenario, you can store the users' information on an external table, and then implement the security through this table. Here is a blog which describes how to implement
    it step by step, please see:
    SSAS – Tabular Dynamic Security
    Hope this helps.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Memory usage in Analysis Services tabular model

    Hello,
    I've been researching and investigating trying to understand what is consuming memory resources in a tabular model that I'm working with. Using SQL Server Management Studio, the Estimated Size of the database is reported as 7768.34 MBs. Using
    Kasper de Jonge's BISM Server Memory Report, the database is reported as 15,465.13 MBs. However, a majority of the fields in the BISM Server Memory Report are empty, and so I cannot determine what is consuming the memory. The data source for this particular
    workbook is $SYSTEM.DISCOVER_OBJECT_MEMORY_USAGE.
    For example: I drill-down to an individual column (ColumnA) in the BISM Server Memory Report (Database > Dimensions > Table > In-Memory Table > Columns > Column) and the reported memory usage is 706.97 MBs. Underneath ColumnA,
    I see a blank level with a reported memory usage of 623.59 MBs and a Segments level with a reported memory usage of 83.39 MBs. Looking at $SYSTEM.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS, if I SUM the USED_SIZE for ColumnA, it totals roughly 83 MBs which matches
    what is reported in the BISM Server Memory Report for the segment size. How do I determine what the other 623.59 MBs is being used for? Again, this discrepancy occurs for all columns in the model and not just this one example.
    Thanks!

    Follow up to my original question. It appears that the blank levels (at least under the column level) the Kasper de Jonge's BISM Server Memory Report reports the dictionary size of the column. The memory usage size matches the DICTIONARY_SIZE attribute in
    the $SYSTEM.DISCOVER_STORAGE_TABLE_COLUMNS. I made a mis-assumption as to what the information $SYSTEM.DISCOVER_STORAGE_TABLE_COLUMN_SEGMENTS was providing.
    In my original post I reference one database in particular in where the Estimated Size property in the Database Properties dialog is listed as 7,768.34 MBs. and the $SYSTEM.DISCOVER_OBJECT_MEMORY_USAGE reports 15,465.13 MBs. Thoughts, comments, or opinions
    on why the Estimated Size property does not match what's reported in $SYSTEM.DISCOVER_OBJECT_MEMORY_USAGE?
    Thanks!

  • Error while importing data in SSAS Tabular Model

    I am new to the concept Tabular Model in SSAS 2014.
    I am trying to create one tabular model based on Adventureworks DW 2014.
    I am getting below error while importing tables to create model.
    "OLE DB or ODBC error: Login failed for user 'ASIAPAC\CSCINDAE732028$'.; 28000.
    A connection could not be made to the data source with the DataSourceID of '98c2d415-1e84-469c-a170-2bcacd779c1a', Name of 'Adventure Works DB from SQL'.
    An error occurred while processing the partition 'Customer_65240c88-55e7-416c-a7ac-732dece8be8e' in table 'Customer_65240c88-55e7-416c-a7ac-732dece8be8e'.
    The current operation was cancelled because another operation in the transaction failed."
    But while creating the Datasource, it has created successfully (below img)
    But while importing the facing the below error.
    Note:
    I have multiple instances in my system with unique names.
    Is this causing any ambiguity issues in selecting right instances?

    Hi Naveen,
    Based on your screenshots, you fail to open a connection to the data source. Right?
    In this scenario, the first screenshot you post is for creating a connection to server with the current windows authentication, not for connecting a data source. So the "Test Connection succeed" means your current windows user can connect
    to server, not connect to the database your selected in the dropdown list. Then you click next, you can choose account to access the data source. Based on the information, your service account "'ASIAPAC\CSCINDAE732028$" doesn't have permission to
    access the database you selected. Please grant the permission for the service account in SSMS.
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou
    If you have any feedback on our support, please click here.

  • Creating Tabular Models using AMO

    Hi,
       I am trying to setup Tabular Models on SSAS using AMO referring to the materials available at Technet and the Codeplex AMO Tabular Tutorial. I have a few questions/issues
    currently and was hoping I could get some help from here. The tool I am working with currently supports Multidimensional mode via AMO and I am trying to extend most of its features from the Multidimensional model. So, some of the questions might not exactly
    fit into the Tabular model and if so, please do point them out.
    RELATIONSHIPS: While creating relationships among tables,
    would it suffice to create the relationship in the Data Source View Schema (as the tool currently does for the Multidimensional mode) or should we explicitly create relationships under the dimensions that we create to represent the table in Tabular mode? The
    latter is the approach mentioned in the tutorial and I wanted to know if there is any difference between the relationships created using the 2 methods.
    MEASURES: I am trying to create measures using Native
    OLAP Measure objects in the Tabular model. I understand the shortcomings of this type of measures but I would still need to create them for basic aggregation functions like SUMMATION, COUNTS etc… I tried creating the measures using the same Measure Group that
    was created to represent the table and used a ColumnBinding to the Measure’s source-column for the measure object’s source property but I get the following error:
    Binding for VertiPaq measure MEASURE_NAME in measure group MEASURE_GROUP_NAME is invalid because it does not match any property binding of the fact dimension
    TABLE_NAME
    Am I missing anything here? Is there a better way of creating OLAP measures in TABULAR model without using the default Table dimensions that we create?
    HIERARCHIES: I tried creating new Dimensions to hold
    hierarchies for a table but when creating them, I get the error about MISSING ROWNUMBER attribute. Is this attribute mandatory for every dimension that is created?
    To avoid this problem, I used the same dimension that was created to represent the table and tried adding inter-table attribute relationships to it but I get the following exception
    message which I cannot figure out.
    VertiPaq property ‘’ cannot have a name binding.
    In general, can we create separate measure groups and dimensions, apart from the ones we use to represent the table, to store the custom measures and hierarchies? Is this a recommended
    approach? This way, I am trying to keep things in parallel with the Multidimensional model that our tool currently supports but when I create such individual dimensions and measure groups, I get an error on the mismatch between the number of measure groups
    and dimensions in the table.
       Please bear with the long list of questions. I could not find any help online for these and so am posting them all here.
    Thank you.

    RELATIONSHIPS: While creating relationships among tables,
    would it suffice to create the relationship in the Data Source View Schema (as the tool currently does for the Multidimensional mode) or should we explicitly create relationships under the dimensions that we create to represent the table in Tabular mode? The
    latter is the approach mentioned in the tutorial and I wanted to know if there is any difference between the relationships created using the 2 methods.
    No - relationships in the DSV have no impact on the end model. You need to explicitly create relationships between your dimensions and measure groups for them to be picked up as relationships in your tabular model.
    MEASURES: I am trying to create measures using Native OLAP
    Measure objects in the Tabular model. I understand the shortcomings of this type of measures but I would still need to create them for basic aggregation functions like SUMMATION, COUNTS etc… I tried creating the measures using the same Measure Group that was
    created to represent the table and used a ColumnBinding to the Measure’s source-column for the measure object’s source property but I get the following error:
    Binding for VertiPaq measure MEASURE_NAME in measure group MEASURE_GROUP_NAME is invalid because it does not match any property binding of the fact dimension TABLE_NAME
    Am I missing anything here? Is there a better way of creating OLAP measures in TABULAR model without using the default Table dimensions that we create?
    There is no concept of native measures in a Tabular model. You need to create all your measures as "Calculated Measures" in AMO, but using the appropriate DAX expressions instead of MDX.
    HIERARCHIES: I tried creating new Dimensions to hold hierarchies
    for a table but when creating them, I get the error about MISSING ROWNUMBER attribute. Is this attribute mandatory for every dimension that is created?
    To avoid this problem, I used the same dimension that was created to represent the table and tried adding inter-table attribute relationships to it but I get the following exception message
    which I cannot figure out.
    VertiPaq property ‘’ cannot have a name binding.
    Yes, I believe every table needs the hidden RowNumber attribute. The
    TableAddEmptyTable function in the tabular AMO sample on codeplex shows you how to create this.
    In general, can we create separate measure groups and dimensions, apart from the ones we use to represent the table, to store the custom measures and hierarchies? Is this a recommended
    approach? This way, I am trying to keep things in parallel with the Multidimensional model that our tool currently supports but when I create such individual dimensions and measure groups, I get an error on the mismatch between the number of measure groups
    and dimensions in the table.
    No you can't create any extra structures. Tabular projects only supports a subset of AMO. You need to follow the example on codeplex very closely and read all the code comments if you are making changes because it's very easy to break things.
    My suggestion is to create an abstraction layer either using the TabularAMO library from codeplex as it is or creating your own library if you only need a subset of the functionality. This will mean that your core code is not too tightly bound to AMO. The
    reason for this is that I would hope that MS will replace AMO with something better for Tabular models in a coming release and having a clear abstraction layer should make it easier to update to a new API.
    http://darren.gosbell.com - please mark correct answers

  • Error in import : tabular model with Oracle

    Hi there,
    I seems to be getting very strange error.
    I am trying to build tabular model over oracle db. I get test successfull for connection, able to validate the query, and in query designer, if i execute it shows me data too. But when i press finish, it gives me error while importing.
    ERROR from import screen
    OLE DB or ODBC error: ORA-01033: ORACLE initialization or shutdown in progress.
    A connection could not be made to the data source with the DataSourceID of '8ffdbfd9-e21f-442d-9040-b66dc074c87d', Name of 'XXXXXXX'.
    An error occurred while processing the partition 'XXX_3fc019fa-4d1a-4d39-9317-3dcde3297579' in table 'XXXXXX'.
    The current operation was cancelled because another operation in the transaction failed.
    I cant figure this out,Even in Execl powerpivot i am able to pull the data in.
    Any idea..TIA
    Rahul Kumar, MCTS, India, http://sqlserversolutions.blogspot.com/

    Hi Rahul,
    Based on my research, for the error “OLE DB or ODBC error: ORA-01033: ORACLE initialization or shutdown in progress.
    A connection could not be made to the data source with the DataSourceID of '8ffdbfd9-e21f-442d-9040-b66dc074c87d', Name of 'XXXXXXX'.”
    It seems that it related to Oracle database, here are some links about how to troubleshoot this issue, please see:
    ERROR - ORA-01033 oracle initialization or shutdown in progress
    ERROR - ORA-01033 oracle initialization or shutdown in progress
    If the issue persists, since I am not an expert on Oracle database, you can post the question on the Oracle forum, so that you can get more help.
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here.
    Charlie Liao
    TechNet Community Support

  • Tabular Modeling. What is the best practice for importing data into VS to limit the records in the designer?

    Should I wrap the queries in a procedure with a @StartDate and @EndDate and create a test partition to pass a small date range? 
    Or can i use the Table properties screen to put the command there and will it run and not be affected or affect the partitions?This would be nice if this SQL statement on this screen was independent of the partitions and I could just leave it with the the
    command text = EXEC TransactionDetail '2014-01-01', '2014-05-31' Especially since if you have many tables that load based on a date range. i would not want to jump in and change that query on all of them.
    Is there a a way to have a parameter in the project so all tables would get the same @startDate and @EndDate so I could change it in one place?
    And I am not stuck to these questions\options, If there is a better way to mass change the queries to run a subset of data for the designer I'd like to hear it.
    Thank You,
    Phil

    Hi Phil,
    According to your description, you are looking for the best way to control the rows that are loaded into a table, right?
    When importing data to a table of tabular model, we can apply filters to control the rows that are loaded into a table. After you have imported the data, you cannot delete individual rows. However, you can apply custom filters to control the way that
    rows are displayed. Rows that do not meet the filtering criteria are hidden. For the detail information about it, please refer to the link below.Filter Data in a Table (SSAS Tabular)
    If I have anything misunderstand, please point it out.
    Regards,
    Charlie Liao
    If you have any feedback on our support, please click
    here
    Charlie Liao
    TechNet Community Support

  • Error during the attach of a Tabular model - SSAS 2012

    Hi,
    I've changed the DataDir for a SSAS Tabular instance by specifying another drive letter and a path like "E:\Microsoft SQL Server\MSAS11.SQLSERVER2012\OLAP\Data".
    After the restarting of the SSAS service I'm trying to attach the tabular model but when I select a folder I've an error about the specified folder is not valid because it doesn't correspond to the naming convention.
    How can I solve this issue, thanks?

    Hi pscorca,
    After you rename the "DataDir" path for SSAS Tabular instance, please copy Tabular database files to the new location(E:\Microsoft SQL Server\MSAS11.SQLSERVER2012\OLAP\Data). Then, please follow the steps below:
    Right-click on the Tabular model instance-> Properties, switch to "General" page.
    Select "Show Advanced(All) Properties" option, and then set the following value for "AllowedBrowsingFolders" property.
    C:\Program Files\Microsoft SQL Server\MSAS11.TABULAR2\OLAP\Backup\|C:\Program Files\Microsoft SQL Server\MSAS11.TABULAR2\OLAP\Log\|C:\Program Files\Microsoft SQL Server\MSAS11.TABULAR2\OLAP\Data\|E:\Microsoft SQL Server\MSAS11.SQLSERVER2012\OLAP\Data
    Restart SSAS Tabular instance.
    When you attach a Tabular database, please select the corresponding database which under "E:\Microsoft SQL Server\MSAS11.SQLSERVER2012\OLAP\Data" path.
    Please let me know if you have any questions.
    Regards,
    Elvis Long
    TechNet Community Support

  • How to create Date relationship with Fact and DimDate on Tabular model?

    I have two tables in SQL Server 2014 SSAS Tabular model.
    DimDate(defined as data table) have date column (format is 1.7.2010 00:00:00) This table have been impoted from AdventureWorks.
    FactSales have SalesDate column (format is 25.03.2015 18:08:05). This table have been imported from Excel. Colums is defined as date in Excel.
    When I set Data Type as date in Tabular I get error:
    "Datatype conversion failed for table...Value:'25.03.2015 18:08:05'"
    I have tried to create calculated colums like =DATEVALUE([SalesDate]) and =DATE(YEAR([SalesDate]),MONTH([SalesDate]), DAY([SalesDate])), but getting error.
    What should I do so that I can create relationship between Date column?
    Kenny_I

    Hi Kenny_l,
    According to your description, you fail to convert the data type into date after creating calculated column with expression. Right?
    In Analysis Services, the tabular will detect the source data type to determine which data type can be converted to. If you want to convert a data into date type, no matter this data is from database or a file, you must make sure this data can be recognized
    as a date in SQL Server. In this scenario, the format '25.03.2015 18:08:05' can't be recognized as date even you apply DAX functions. So please change the format of the column in excel like "03.25.2015 18:08:05".
    PS: For testing, you can create a temp table and insert sample text into a date column, because the date format in SQL Server depends on the location selection when installing.
    After retrieving data from data source and all values within column are correct format, you can change the data type into Date, and select the format you expect.
    If you have any question, please feel free to ask.
    Best Regards,
    Simon Hou
    TechNet Community Support

  • Tabular - Model vs Perspective

    Hi,
    I have a tabular model which has been in production for over a year with 30-some tables and reports in the 100's. Now I have been asked to add 10-15 new columns in the tabular model but they should only be available to a handful of people. Up until now I
    have not had any Perspectives but I thought now is probably the time to add it. New columns are not "super sensitive" but I have been asked that they should not be available for everyone. 
    My question is this: is it possible to restrict access to these columns (hide them) in the Model "perspective" and show them in the new FullAccess perspective? If I choose "hide from client tools" the columns are not visible in any of
    the perspectives and it seems I can't hide them in the default perspective in the perspective designer...
    Lumbago
    www.thefirstsql.com

    Hi Lumbago,
    According to your description, you added 10-15 new columns in the tabular model but they should only be available to a handful of people, right?
    To secure data, you can define security roles. Roles can limit viewable metadata and data to only those objects defined in the role. Here is a blog which describes how to implement Column Security with #SSAS Tabular and #DAX, please see:
    http://geekswithblogs.net/darrengosbell/archive/2014/04/22/implementing-column-security-with-ssas-tabular-and-dax.aspx
    Regards,
    Charlie Liao
    TechNet Community Support

  • Processing Tabular Model

    1  We are developing a Tabular Model. In this model there are likely to be about 15 dimension tables and 2 main fact tables.
    2  One of the fact tables is very large approximately 70 million rows and will increase to 150 million rows in about 2-3 years time.
    3  So on the cube we have created partitions and are using the incremental processing. What we have done is to create a partition definition table header and lines. At the header level we will store the name of the measure group on which we wish to
    create partitions, and in the lines table we will create the definition of each partition. Using an sp we will mark of those rows of the partition lines which we wish to reprocess. Such partitions will be dropped and recreated. So far this is working well.
    4  I want to generalize this solution so that I works across different projects without any changes.
    Now I have two questions :
    Question 1 :
    If I make changes in the tabular project and deploy the same, I believe all partitions will get deleted and all the data will need to be pulled in again. This will happen even if I add a calculated measure. Is there any method to overcome this ?
    Question 2 :
    What is the mechanism of only processing certain measure tables incrementally and all other tables fully ? In my above example only one table has partitions. So if I want to process only the current partition of that table, and all other tables how do I
    achieve this ?
    Sanjay Shah
    Prosys InfoTech, Pune, India

    1) if you only add a measure or a calculated column, you do not need to read data from data source. If you have problem with deployment within VS, consider using Deployment Wizard.
    2) A complete description of process strategies is included in a chapter of our book (http://www.sqlbi.com/books/microsoft-sql-server-2012-analysis-services-the-bism-tabular-model).
    In general, you can control which partition/tables you want to process and in which way, using XMLA scripts, PowerShell and other tools. The easiest way to create an XMLA script is using the Generate Script feature in SSMS when you use the process wizard.
    Marco Russo http://ssasworkshop.com http://www.sqlbi.com http://sqlblog.com/blogs/marco_russo

  • Rename Fact dimension in tabular model

    Hi there,
    I have a small Question I wanna rename Fact dimension  in tabular model I mean I was using Fact table as a dimension as well.
    Fact Table: SaleOrder
    Dimension: SalesOrder
    So under measures it should display as Fact_SalesOrder and Under dimensions it should display as Dim_SalesOrder.
    where I can rename this ? at the moment in the two places it is displaying as
    SalesOrder
    Thanks for your help

    Hi Siva,
    In your sceanrio, you can create a view on the fact table, and then import this view to tabular model as the dimension table, so measure group name with not duplicated with the dimension name. Besides, there are benefits in creating the cube from database
    views that sits on top of the underlying fact and dimension tables. Please refer to the link below to see the detail information about it.
    Benefits of Creating SSAS Cube from Database Views
    Hope this helps.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Analysis Services Tabular Model 2014 vs 2012

    Hi,
    A very simple question : with the same tabular model is 2014 faster than 2012 from query point of view (is the Storage Engine or Formula Engine faster than the previous version ) ?
    Thanks 

    Hi MIB1,
    According to your description, you want to know if the SSAS 2014 tabular model has better performance than SSAS 2012 tabular. Right?
    With exception to added functionality supporting Power View Reports against Multidimensional Models, SQL Server 2014 Analysis Services is unchanged from the previous release. There's nothing improvement on SSAS tabular.
    Reference:
    What's New in Analysis Services and Business Intelligence
    If you have any question, please feel free to ask.
    Regards, 
    Simon Hou
    TechNet Community Support

  • Migrated data from Powerpivot to SSAS tabular model, can i change the Excel source to SSAS?

    Hi
    I have developed some powerview reports in  Excel/powerpivot files. Now, i have migrated powerpivot data from these files to SSAS tabular cubes. The question is if i can change the source of the Excel files, to "feed" my powerview reports
    with these SSAS Tabular cubes, or do i have to create my powerview reports again via new Excel files or PowerBI designer (and "point" these to SSAS Tabular cubes). It totally seems not logical and so i think that's a good question.
    Kind regards

    Hi Pascal,
    Power View only supports Microsoft BI Semantic Model (BISM) for Power View data sources. BISM files are designed specifically for connecting to Excel workbooks and SSAS tabular models. The Excel workbooks can be in the PowerPivot Gallery on a SharePoint
    site or in a SharePoint library. In this issue, you can publish the PowerPivot workbooks to a SharePoint library or the PowerPivot Gallery, and then create RSDS file or BISM data source to the workbooks. Here are the connection string samples:
    An Excel workbook in Shared Documents on a SharePoint site:
    Data Source=”http://<SharePointSite>/Shared Documents/MyExcelWorkbook.xlsx”
    An Excel workbook in the PowerPivot Gallery on a SharePoint site:
    Data Source=”http://<SharePointSite>/PowerPivot Gallery/MyExcelWorkbook.xlsx”
    Reference:
    http://office.microsoft.com/en-us/excel-help/create-a-connection-to-a-data-model-for-power-view-HA102835737.aspx?CTT=5&origin=HA102834736
    Regards, 
    Mike Yin
    TechNet Community Support

Maybe you are looking for

  • New Infinity 2 order HELP!!

    hey there, I have just ordered BT Infinity 2 fibre with phone (evening and weekends unlimited) and i'd just like to know the process as I have had an email to say my line will be swapped from Sky on 3rd Feburary 2014 and THEN Infinity 2 will be order

  • Problem in import Purchase Order GRN cancellation

    Dear Experts, I have a problem in related to import PO. I have created an Import PO and made GRN, but due to the exchange rate not fixed at the time of PO creation, I got some difference in amount in local currency Vs Ord. currency at the time of GRN

  • Updation of Table EDPAR

    Hi all , Can anybody tell how table EDPAR can be updated or changed any transaction for that ?? Thanks in Advance , Vivek K

  • Flash CS6

    As I know adobe flash cs6 isn't an apple product, I had a question and I was also wondering if it had to do with my mac? I hope someone can answer my question- When I opened flash and went to click a color from the swatches, I noticed that when I clc

  • Logic Pro X: Audio file not found?

    I have Logic Pro X, and this problem has never occurred before but... I was mixing one of my songs and everything was perfect, then I made sure to save it. As I reopened literally 10 seconds later (not sure why I even closed it) it was showing me a m