Use of Loading Group in TLB to sort product

Hi Experts,
Currently we are using Loading Group in Product Master to prioritize Product for TLB. I would like to know about sorting logic where Loading Group is same for multiple products. How system does sort in this type of scenario?
E.g.
Sr.No.     Product     Loading Group
1     10002577     AC80
2     10002813     AE60
3     10002815     AE60
4     10002816     AE60
5     10002817     AE60
6     10002869     AI60
7     10000505     AI60
8     10003318     AI60
9     10002872     AI60
10     10002873     AI60
When TLB sort product, it will give first priority to AC80 product, then AE60 & then AI60 but how system sort products which falls in AE60 OR AI60? Which criteria system use to sort products if loading group is same?
Can you please help me out to understand the sorting logic?
Thanks & regards
Sujay Joshi

Hi Sujay,
The TLB sorts the loading groups and then the products within the loading group alphabetically.
This is the reason why when TLB sorts the products, it gives first priority to AC80 product, then AE60 & then AI60 .
Regards,
Nikhil

Similar Messages

  • TLB ignoring Loading Group

    Folks,
    I am running TLB for products which have different loading group ( Loading group is maintained in Location Product under GR/GI Tab). I am using Straight Loading for TLB run. My understanding of TLB and loading group in straight loading imethod is system will not club products with different loading group into one TLB order.However after TLB run i see that system has created TLB orders which has multiple products ( i.e different loading group products are clubbed in one TLB order).
    For example let us assume that we have products A, B and C. Product A has loading Group 1, B has loading group 2 and C has loading Group 3. After TLB run , i see product A and B in one TLB order, which should not be the case i feel.
    Is my understanding of TLB and loading group correct? Any inputs on the above issue will be highly appreciated.
    Thanks in advance !!
    Regards,
    Pawan.

    Hi,
    1) How is system going to give priority to Loading Groups? I do not think there is any setting available which tells the product to give priority to loading group? Alternatively is it possible that sytem may give priority to loading group based on alphabetical order.
    Ans: Iam not sure about alphabetical but if you are using the numerical then it follow as said before. First it load for 0000 then 0001 etc.. If Truck has allowed space then it try to load from the other loading group. In our senario we are using it and system is working as i said.
    2) if even straight loading is going to club products, then i think there is no difference between straight loading and load balancing.
    Ans: In load balancing it will not check Loding group but in strait loading it checks loding group. I
    3)Many a times there is a business requirement that few material ( esp hazardous material) should not be transported with other materials. Then in those cases how TLB handles those materials.
    Ans: You can achieve this requirement by maintaining differe nt MOT's (without all product flag) and maintain TLB profile differently for the hazardous material and maintain product specific MOT for those materials.
    Hope this helps.
    Regards,
    Kishore Reddy.

  • Updating Loading Group on plant record using BAPI

    I am using BAPI to extend to a plant
    BAPI_MATERIAL_SAVEDATA
    I am passing                     
      plantdata-loadinggrp   = wa_plantdata-loadinggrp.
      plantdatax-loadinggrp  = 'X'.
      plantdata-plant        = p_p_plant.
      plantdatax-plant       = p_p_plant.
    At the time of the BAPI call the data is in the table
    All other plant data gets created except for this loading group field (Sales/general plant)  and only when I am usign a specific plant.   It works fine for other plants
    Anybody know any config or setting related to this ?

    solved it

  • Fiori Launchpad Error: Failure - Unable to load groups Resolved!!!!

    Hey Good Fellas,
    This is a solution rather than a discussion, which i thought to do to with a kind generosity of spirit to give back to the community.
    It gives me immense pleasure and happiness to write this particular post, because it is through SCN only i have been able to find the solution.
    Ok, let's get down to the Business quickly, shall we?
    I have been configuring Fiori App - Approve Purchase Orders. All the tasks for this have been done from SAP Help Site.
    Recently I upgraded the ERP/Gateway (Its an embedded deployment) to EHP 7 and UI to SP09, post which the Fiori Launchpad started giving the below error.
    If you look around the SCN, you will probably hit across these threads and so did I.
    http://scn.sap.com/thread/3571448
    SAP Fiori - UI Add-on SP09 update troubleshooting
    With the first thread, I tried everyone's solution and suggestions which had worked for others, but it didn't for me.
    With second thread (Excellently posted by Masayuki Sekihara), it started off with lots of efforts and double-checks applying all of these notes hoping that it would resolve the issue. But I had done all the steps 1-8 and still didn't get the solution.
    Besides the above two mentions I did umpteen number of other ridiculous things like implementing this note, that note, De-implementing another of the notes and everything i found as possible solutions on the Web.
    Now, the breakthrough here is checking the errors in /IWFND/ERROR_LOG. Well, on the browser it used to give error 'Failure - Unable to Load Groups' while at the server side, actual error was something unusual and remotely not related to what the error looked like in the browser. In the back-end, I was getting the error "Data Provider implementation 'ZINTEROP_MODEL_0001_BE' '01' 'DEFAULT' does not exist". And this is something i know how to correct.
    So just keep looking for the errors in /IWFND/ERROR_LOG and shift your focus there. That's really a good place to trouble-shoot.
    And for the record to take care of the error 'Data Provider implementation 'ZINTEROP_MODEL_0001_BE' '01' 'DEFAULT' does not exist' you have to assign the DPC (Data Provider Class) to DMC (Data Modelling Class) correctly by navigating to IMG - SAP Netweaver - Gateway - Former Development - Local OData Channel Development - Registration - Assign Data Provider to Data Model. There you define the Data provider with the correct Software Version and Voila you are all set! You don't have to worry about the duplicate entry with another Software version, your correct software version is the one you have defined in the System Alias which you are using inside the service, which is generally DEFAULT.
    I am able to get inside the Fiori Launchpad now (I was quite at edge to really give up as all the possible solutions had been ruled out for me ) and do see the tiles of the apps configured.
    I hope this is going to be a good take-away for others who are having Odata handling issues.
    The floor is all yours now for your kind feedback. Be gentle with me, I am a very shy blogger.
    Best Regards,
    Akshay

    Steve,
    If you still have the error, this is what I did to resolve mines.
    1) Goto SPRO - SAP Netweaver - Gateway - Former Development - Local OData Channel Development - Registration - Assign Data Provider to Data Model
    2) Find out the entry for 'ZINTEROP_MODEL_0001_BE'
    3) Copy that entire row in a notepad for all the fields
    4) Create new Entry with all the fields, except change the Software Version to DEFAULT (As that's what missing for you'
    5) Save and give that another try.
    This is what I exactly wrote in my solution as well.
    Akshay wrote:
    And for the record to take care of the error 'Data Provider implementation 'ZINTEROP_MODEL_0001_BE' '01' 'DEFAULT' does not exist' you have to assign the DPC (Data Provider Class) to DMC (Data Modelling Class) correctly by navigating to IMG - SAP Netweaver - Gateway - Former Development - Local OData Channel Development - Registration - Assign Data Provider to Data Model. There you define the Data provider with the correct Software Version and Voila you are all set! You don't have to worry about the duplicate entry with another Software version, your correct software version is the one you have defined in the System Alias which you are using inside the service, which is generally DEFAULT.
    A screen-shot perhaps helps you more.
    Let the other duplicate entry with Software Version other than Default be there. Important is that the software version that you have defined in your alias must be the one over here.
    Let me know how that goes?
    -Akshay

  • Delivery split for STO based on Material Loading Group

    Hello, I want to split the Delivery created from STO based on material loading group. I checked SAP Note:166397 and it requires copying routine. But this solution is not accepted and looking for other solutions. Please help with any user-exit or badi which can solve the purpose. Delivery needs to be created from VL10B transaction.
    Edited by: ananthula sumanth on Aug 6, 2010 9:12 AM

    The copy routine is designed for this purpose - it's not clear why the client does not allow the use of it. In my opinion you should find the root casue of their resistance and offer an acceptable solution (if it is possible).
    If you want to solve it in another way I guess you should go into a big development and develop your own copy routine functionality which could mean much more work (for you) and money (of your client) - instread of modifying a copy routine (maybe within few minutes).

  • Splitting Purchase Order based on Loading Group.

    Hi,
    My client is a fashion retailer. They have two Loading Groups, General Merchendise that comes in cartons and Hung garments that com hung on hangers.
    PO`s are created for every Vendor by Follow-on processing of allocation or by replenishment run.
    It is desired to split our Purchase Orders based on loading group.  So that there is a separate PO for General Merchendise and Hung Garments. These can then be handled separately during GR and furthur processing.
    Would like to know how to configure this.
    Thanks
    Barry

    You need to Define Two PO document types one for General Merchendise and other for Hung garments
    and use BADI  ME_PROCESS_PO_CUST  and incorporate the Logic that
    for PO Type General Merchendise the Loading Group of Hung garments Not allowed and Similarly
    for PO type Hung garments the Loading group of General Merchendise not allowed ..

  • SharePoint 2010 List View with Group By is Not Sorting Correctly

    My organization has a list in SharePoint 2010 that is being used as a FAQ. The list contains the following columns:
    Question-  Single line of text 
    Answer-  Multiple lines of text 
    Priority-  Number 
    Priority is the importance of the question and is a unique integer number. Higher Priority questions are to be displayed at the top of the list.
    A view has been created to display the Question and Answer columns, grouped by Question, and sorted by Prioity. The problem is that the SharePoint is not sorting by priority. I have added display of Priority and embedded a screen capture showing the order
    going non sequentially from 1,005 to 1,024 to 1,015.
    Is this a known defect? Is there a workaround? Are there other options?
    Regards,
    Darian.

    Hi,
    Based on your description, I have done a test. I can reproduce your issue.
    Per my test, I come up with a conclusion.
    When we group by Question, the list view is be grouped. Then we sort by Priority, actually it is only sort the item within a group rather than the entire list.
    So the items inside of different groups out of order is normal.
    Best Regards,
    Lisa ChenForum Support
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected]
    Lisa Chen
    TechNet Community Support

  • Sliding Window Table Partitioning Problems with RANGE RIGHT, SPLIT, MERGE using Multiple File Groups

    There is misleading information in two system views (sys.data_spaces & sys.destination_data_spaces) about the physical location of data after a partitioning MERGE and before an INDEX REBUILD operation on a partitioned table. In SQL Server 2012 SP1 CU6,
    the script below (SQLCMD mode, set DataDrive  & LogDrive variables  for the runtime environment) will create a test database with file groups and files to support a partitioned table. The partition function and scheme spread the test data across
    4 files groups, an empty partition, file group and file are maintained at the start and end of the range. A problem occurs after the SWITCH and MERGE RANGE operations, the views sys.data_spaces & sys.destination_data_spaces show the logical, not the physical,
    location of data.
    --=================================================================================
    -- PartitionLabSetup_RangeRight.sql
    -- 001. Create test database
    -- 002. Add file groups and files
    -- 003. Create partition function and schema
    -- 004. Create and populate a test table
    --=================================================================================
    USE [master]
    GO
    -- 001 - Create Test Database
    :SETVAR DataDrive "D:\SQL\Data\"
    :SETVAR LogDrive "D:\SQL\Logs\"
    :SETVAR DatabaseName "workspace"
    :SETVAR TableName "TestTable"
    -- Drop if exists and create Database
    IF DATABASEPROPERTYEX(N'$(databasename)','Status') IS NOT NULL
    BEGIN
    ALTER DATABASE $(DatabaseName) SET SINGLE_USER WITH ROLLBACK IMMEDIATE
    DROP DATABASE $(DatabaseName)
    END
    CREATE DATABASE $(DatabaseName)
    ON
    ( NAME = $(DatabaseName)_data,
    FILENAME = N'$(DataDrive)$(DatabaseName)_data.mdf',
    SIZE = 10,
    MAXSIZE = 500,
    FILEGROWTH = 5 )
    LOG ON
    ( NAME = $(DatabaseName)_log,
    FILENAME = N'$(LogDrive)$(DatabaseName).ldf',
    SIZE = 5MB,
    MAXSIZE = 5000MB,
    FILEGROWTH = 5MB ) ;
    GO
    -- 002. Add file groups and files
    --:SETVAR DatabaseName "workspace"
    --:SETVAR TableName "TestTable"
    --:SETVAR DataDrive "D:\SQL\Data\"
    --:SETVAR LogDrive "D:\SQL\Logs\"
    DECLARE @nSQL NVARCHAR(2000) ;
    DECLARE @x INT = 1;
    WHILE @x <= 6
    BEGIN
    SELECT @nSQL =
    'ALTER DATABASE $(DatabaseName)
    ADD FILEGROUP $(TableName)_fg' + RTRIM(CAST(@x AS CHAR(5))) + ';
    ALTER DATABASE $(DatabaseName)
    ADD FILE
    NAME= ''$(TableName)_f' + CAST(@x AS CHAR(5)) + ''',
    FILENAME = ''$(DataDrive)\$(TableName)_f' + RTRIM(CAST(@x AS CHAR(5))) + '.ndf''
    TO FILEGROUP $(TableName)_fg' + RTRIM(CAST(@x AS CHAR(5))) + ';'
    EXEC sp_executeSQL @nSQL;
    SET @x = @x + 1;
    END
    -- 003. Create partition function and schema
    --:SETVAR TableName "TestTable"
    --:SETVAR DatabaseName "workspace"
    USE $(DatabaseName);
    CREATE PARTITION FUNCTION $(TableName)_func (int)
    AS RANGE RIGHT FOR VALUES
    0,
    15,
    30,
    45,
    60
    CREATE PARTITION SCHEME $(TableName)_scheme
    AS
    PARTITION $(TableName)_func
    TO
    $(TableName)_fg1,
    $(TableName)_fg2,
    $(TableName)_fg3,
    $(TableName)_fg4,
    $(TableName)_fg5,
    $(TableName)_fg6
    -- Create TestTable
    --:SETVAR TableName "TestTable"
    --:SETVAR BackupDrive "D:\SQL\Backups\"
    --:SETVAR DatabaseName "workspace"
    CREATE TABLE [dbo].$(TableName)(
    [Partition_PK] [int] NOT NULL,
    [GUID_PK] [uniqueidentifier] NOT NULL,
    [CreateDate] [datetime] NULL,
    [CreateServer] [nvarchar](50) NULL,
    [RandomNbr] [int] NULL,
    CONSTRAINT [PK_$(TableName)] PRIMARY KEY CLUSTERED
    [Partition_PK] ASC,
    [GUID_PK] ASC
    ) ON $(TableName)_scheme(Partition_PK)
    ) ON $(TableName)_scheme(Partition_PK)
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_GUID_PK] DEFAULT (newid()) FOR [GUID_PK]
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_CreateDate] DEFAULT (getdate()) FOR [CreateDate]
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_CreateServer] DEFAULT (@@servername) FOR [CreateServer]
    -- 004. Create and populate a test table
    -- Load TestTable Data - Seconds 0-59 are used as the Partitoning Key
    --:SETVAR TableName "TestTable"
    SET NOCOUNT ON;
    DECLARE @Now DATETIME = GETDATE()
    WHILE @Now > DATEADD(minute,-1,GETDATE())
    BEGIN
    INSERT INTO [dbo].$(TableName)
    ([Partition_PK]
    ,[RandomNbr])
    VALUES
    DATEPART(second,GETDATE())
    ,ROUND((RAND() * 100),0)
    END
    -- Confirm table partitioning - http://lextonr.wordpress.com/tag/sys-destination_data_spaces/
    SELECT
    N'DatabaseName' = DB_NAME()
    , N'SchemaName' = s.name
    , N'TableName' = o.name
    , N'IndexName' = i.name
    , N'IndexType' = i.type_desc
    , N'PartitionScheme' = ps.name
    , N'DataSpaceName' = ds.name
    , N'DataSpaceType' = ds.type_desc
    , N'PartitionFunction' = pf.name
    , N'PartitionNumber' = dds.destination_id
    , N'BoundaryValue' = prv.value
    , N'RightBoundary' = pf.boundary_value_on_right
    , N'PartitionFileGroup' = ds2.name
    , N'RowsOfData' = p.[rows]
    FROM
    sys.objects AS o
    INNER JOIN sys.schemas AS s
    ON o.[schema_id] = s.[schema_id]
    INNER JOIN sys.partitions AS p
    ON o.[object_id] = p.[object_id]
    INNER JOIN sys.indexes AS i
    ON p.[object_id] = i.[object_id]
    AND p.index_id = i.index_id
    INNER JOIN sys.data_spaces AS ds
    ON i.data_space_id = ds.data_space_id
    INNER JOIN sys.partition_schemes AS ps
    ON ds.data_space_id = ps.data_space_id
    INNER JOIN sys.partition_functions AS pf
    ON ps.function_id = pf.function_id
    LEFT OUTER JOIN sys.partition_range_values AS prv
    ON pf.function_id = prv.function_id
    AND p.partition_number = prv.boundary_id
    LEFT OUTER JOIN sys.destination_data_spaces AS dds
    ON ps.data_space_id = dds.partition_scheme_id
    AND p.partition_number = dds.destination_id
    LEFT OUTER JOIN sys.data_spaces AS ds2
    ON dds.data_space_id = ds2.data_space_id
    ORDER BY
    DatabaseName
    ,SchemaName
    ,TableName
    ,IndexName
    ,PartitionNumber
    --=================================================================================
    -- SECTION 2 - SWITCH OUT
    -- 001 - Create TestTableOut
    -- 002 - Switch out partition in range 0-14
    -- 003 - Merge range 0 -29
    -- 001. TestTableOut
    :SETVAR TableName "TestTable"
    IF OBJECT_ID('dbo.$(TableName)Out') IS NOT NULL
    DROP TABLE [dbo].[$(TableName)Out]
    CREATE TABLE [dbo].[$(TableName)Out](
    [Partition_PK] [int] NOT NULL,
    [GUID_PK] [uniqueidentifier] NOT NULL,
    [CreateDate] [datetime] NULL,
    [CreateServer] [nvarchar](50) NULL,
    [RandomNbr] [int] NULL,
    CONSTRAINT [PK_$(TableName)Out] PRIMARY KEY CLUSTERED
    [Partition_PK] ASC,
    [GUID_PK] ASC
    ) ON $(TableName)_fg2;
    GO
    -- 002 - Switch out partition in range 0-14
    --:SETVAR TableName "TestTable"
    ALTER TABLE dbo.$(TableName)
    SWITCH PARTITION 2 TO dbo.$(TableName)Out;
    -- 003 - Merge range 0 - 29
    --:SETVAR TableName "TestTable"
    ALTER PARTITION FUNCTION $(TableName)_func()
    MERGE RANGE (15);
    -- Confirm table partitioning
    -- Original source of this query - http://lextonr.wordpress.com/tag/sys-destination_data_spaces/
    SELECT
    N'DatabaseName' = DB_NAME()
    , N'SchemaName' = s.name
    , N'TableName' = o.name
    , N'IndexName' = i.name
    , N'IndexType' = i.type_desc
    , N'PartitionScheme' = ps.name
    , N'DataSpaceName' = ds.name
    , N'DataSpaceType' = ds.type_desc
    , N'PartitionFunction' = pf.name
    , N'PartitionNumber' = dds.destination_id
    , N'BoundaryValue' = prv.value
    , N'RightBoundary' = pf.boundary_value_on_right
    , N'PartitionFileGroup' = ds2.name
    , N'RowsOfData' = p.[rows]
    FROM
    sys.objects AS o
    INNER JOIN sys.schemas AS s
    ON o.[schema_id] = s.[schema_id]
    INNER JOIN sys.partitions AS p
    ON o.[object_id] = p.[object_id]
    INNER JOIN sys.indexes AS i
    ON p.[object_id] = i.[object_id]
    AND p.index_id = i.index_id
    INNER JOIN sys.data_spaces AS ds
    ON i.data_space_id = ds.data_space_id
    INNER JOIN sys.partition_schemes AS ps
    ON ds.data_space_id = ps.data_space_id
    INNER JOIN sys.partition_functions AS pf
    ON ps.function_id = pf.function_id
    LEFT OUTER JOIN sys.partition_range_values AS prv
    ON pf.function_id = prv.function_id
    AND p.partition_number = prv.boundary_id
    LEFT OUTER JOIN sys.destination_data_spaces AS dds
    ON ps.data_space_id = dds.partition_scheme_id
    AND p.partition_number = dds.destination_id
    LEFT OUTER JOIN sys.data_spaces AS ds2
    ON dds.data_space_id = ds2.data_space_id
    ORDER BY
    DatabaseName
    ,SchemaName
    ,TableName
    ,IndexName
    ,PartitionNumber  
    The table below shows the results of the ‘Confirm Table Partitioning’ query, before and after the MERGE.
    The T-SQL code below illustrates the problem.
    -- PartitionLab_RangeRight
    USE workspace;
    DROP TABLE dbo.TestTableOut;
    USE master;
    ALTER DATABASE workspace
    REMOVE FILE TestTable_f3 ;
    -- ERROR
    --Msg 5042, Level 16, State 1, Line 1
    --The file 'TestTable_f3 ' cannot be removed because it is not empty.
    ALTER DATABASE workspace
    REMOVE FILE TestTable_f2 ;
    -- Works surprisingly!!
    use workspace;
    ALTER INDEX [PK_TestTable] ON [dbo].[TestTable] REBUILD PARTITION = 2;
    --Msg 622, Level 16, State 3, Line 2
    --The filegroup "TestTable_fg2" has no files assigned to it. Tables, indexes, text columns, ntext columns, and image columns cannot be populated on this filegroup until a file is added.
    --The statement has been terminated.
    If you run ALTER INDEX REBUILD before trying to remove files from File Group 3, it works. Rerun the database setup script then the code below.
    -- RANGE RIGHT
    -- Rerun PartitionLabSetup_RangeRight.sql before the code below
    USE workspace;
    DROP TABLE dbo.TestTableOut;
    ALTER INDEX [PK_TestTable] ON [dbo].[TestTable] REBUILD PARTITION = 2;
    USE master;
    ALTER DATABASE workspace
    REMOVE FILE TestTable_f3;
    -- Works as expected!!
    The file in File Group 2 appears to contain data but it can be dropped. Although the system views are reporting the data in File Group 2, it still physically resides in File Group 3 and isn’t moved until the index is rebuilt. The RANGE RIGHT function means
    the left file group (File Group 2) is retained when splitting ranges.
    RANGE LEFT would have retained the data in File Group 3 where it already resided, no INDEX REBUILD is necessary to effectively complete the MERGE operation. The script below implements the same partitioning strategy (data distribution between partitions)
    on the test table but uses different boundary definitions and RANGE LEFT.
    --=================================================================================
    -- PartitionLabSetup_RangeLeft.sql
    -- 001. Create test database
    -- 002. Add file groups and files
    -- 003. Create partition function and schema
    -- 004. Create and populate a test table
    --=================================================================================
    USE [master]
    GO
    -- 001 - Create Test Database
    :SETVAR DataDrive "D:\SQL\Data\"
    :SETVAR LogDrive "D:\SQL\Logs\"
    :SETVAR DatabaseName "workspace"
    :SETVAR TableName "TestTable"
    -- Drop if exists and create Database
    IF DATABASEPROPERTYEX(N'$(databasename)','Status') IS NOT NULL
    BEGIN
    ALTER DATABASE $(DatabaseName) SET SINGLE_USER WITH ROLLBACK IMMEDIATE
    DROP DATABASE $(DatabaseName)
    END
    CREATE DATABASE $(DatabaseName)
    ON
    ( NAME = $(DatabaseName)_data,
    FILENAME = N'$(DataDrive)$(DatabaseName)_data.mdf',
    SIZE = 10,
    MAXSIZE = 500,
    FILEGROWTH = 5 )
    LOG ON
    ( NAME = $(DatabaseName)_log,
    FILENAME = N'$(LogDrive)$(DatabaseName).ldf',
    SIZE = 5MB,
    MAXSIZE = 5000MB,
    FILEGROWTH = 5MB ) ;
    GO
    -- 002. Add file groups and files
    --:SETVAR DatabaseName "workspace"
    --:SETVAR TableName "TestTable"
    --:SETVAR DataDrive "D:\SQL\Data\"
    --:SETVAR LogDrive "D:\SQL\Logs\"
    DECLARE @nSQL NVARCHAR(2000) ;
    DECLARE @x INT = 1;
    WHILE @x <= 6
    BEGIN
    SELECT @nSQL =
    'ALTER DATABASE $(DatabaseName)
    ADD FILEGROUP $(TableName)_fg' + RTRIM(CAST(@x AS CHAR(5))) + ';
    ALTER DATABASE $(DatabaseName)
    ADD FILE
    NAME= ''$(TableName)_f' + CAST(@x AS CHAR(5)) + ''',
    FILENAME = ''$(DataDrive)\$(TableName)_f' + RTRIM(CAST(@x AS CHAR(5))) + '.ndf''
    TO FILEGROUP $(TableName)_fg' + RTRIM(CAST(@x AS CHAR(5))) + ';'
    EXEC sp_executeSQL @nSQL;
    SET @x = @x + 1;
    END
    -- 003. Create partition function and schema
    --:SETVAR TableName "TestTable"
    --:SETVAR DatabaseName "workspace"
    USE $(DatabaseName);
    CREATE PARTITION FUNCTION $(TableName)_func (int)
    AS RANGE LEFT FOR VALUES
    -1,
    14,
    29,
    44,
    59
    CREATE PARTITION SCHEME $(TableName)_scheme
    AS
    PARTITION $(TableName)_func
    TO
    $(TableName)_fg1,
    $(TableName)_fg2,
    $(TableName)_fg3,
    $(TableName)_fg4,
    $(TableName)_fg5,
    $(TableName)_fg6
    -- Create TestTable
    --:SETVAR TableName "TestTable"
    --:SETVAR BackupDrive "D:\SQL\Backups\"
    --:SETVAR DatabaseName "workspace"
    CREATE TABLE [dbo].$(TableName)(
    [Partition_PK] [int] NOT NULL,
    [GUID_PK] [uniqueidentifier] NOT NULL,
    [CreateDate] [datetime] NULL,
    [CreateServer] [nvarchar](50) NULL,
    [RandomNbr] [int] NULL,
    CONSTRAINT [PK_$(TableName)] PRIMARY KEY CLUSTERED
    [Partition_PK] ASC,
    [GUID_PK] ASC
    ) ON $(TableName)_scheme(Partition_PK)
    ) ON $(TableName)_scheme(Partition_PK)
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_GUID_PK] DEFAULT (newid()) FOR [GUID_PK]
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_CreateDate] DEFAULT (getdate()) FOR [CreateDate]
    ALTER TABLE [dbo].$(TableName) ADD CONSTRAINT [DF_$(TableName)_CreateServer] DEFAULT (@@servername) FOR [CreateServer]
    -- 004. Create and populate a test table
    -- Load TestTable Data - Seconds 0-59 are used as the Partitoning Key
    --:SETVAR TableName "TestTable"
    SET NOCOUNT ON;
    DECLARE @Now DATETIME = GETDATE()
    WHILE @Now > DATEADD(minute,-1,GETDATE())
    BEGIN
    INSERT INTO [dbo].$(TableName)
    ([Partition_PK]
    ,[RandomNbr])
    VALUES
    DATEPART(second,GETDATE())
    ,ROUND((RAND() * 100),0)
    END
    -- Confirm table partitioning - http://lextonr.wordpress.com/tag/sys-destination_data_spaces/
    SELECT
    N'DatabaseName' = DB_NAME()
    , N'SchemaName' = s.name
    , N'TableName' = o.name
    , N'IndexName' = i.name
    , N'IndexType' = i.type_desc
    , N'PartitionScheme' = ps.name
    , N'DataSpaceName' = ds.name
    , N'DataSpaceType' = ds.type_desc
    , N'PartitionFunction' = pf.name
    , N'PartitionNumber' = dds.destination_id
    , N'BoundaryValue' = prv.value
    , N'RightBoundary' = pf.boundary_value_on_right
    , N'PartitionFileGroup' = ds2.name
    , N'RowsOfData' = p.[rows]
    FROM
    sys.objects AS o
    INNER JOIN sys.schemas AS s
    ON o.[schema_id] = s.[schema_id]
    INNER JOIN sys.partitions AS p
    ON o.[object_id] = p.[object_id]
    INNER JOIN sys.indexes AS i
    ON p.[object_id] = i.[object_id]
    AND p.index_id = i.index_id
    INNER JOIN sys.data_spaces AS ds
    ON i.data_space_id = ds.data_space_id
    INNER JOIN sys.partition_schemes AS ps
    ON ds.data_space_id = ps.data_space_id
    INNER JOIN sys.partition_functions AS pf
    ON ps.function_id = pf.function_id
    LEFT OUTER JOIN sys.partition_range_values AS prv
    ON pf.function_id = prv.function_id
    AND p.partition_number = prv.boundary_id
    LEFT OUTER JOIN sys.destination_data_spaces AS dds
    ON ps.data_space_id = dds.partition_scheme_id
    AND p.partition_number = dds.destination_id
    LEFT OUTER JOIN sys.data_spaces AS ds2
    ON dds.data_space_id = ds2.data_space_id
    ORDER BY
    DatabaseName
    ,SchemaName
    ,TableName
    ,IndexName
    ,PartitionNumber
    --=================================================================================
    -- SECTION 2 - SWITCH OUT
    -- 001 - Create TestTableOut
    -- 002 - Switch out partition in range 0-14
    -- 003 - Merge range 0 -29
    -- 001. TestTableOut
    :SETVAR TableName "TestTable"
    IF OBJECT_ID('dbo.$(TableName)Out') IS NOT NULL
    DROP TABLE [dbo].[$(TableName)Out]
    CREATE TABLE [dbo].[$(TableName)Out](
    [Partition_PK] [int] NOT NULL,
    [GUID_PK] [uniqueidentifier] NOT NULL,
    [CreateDate] [datetime] NULL,
    [CreateServer] [nvarchar](50) NULL,
    [RandomNbr] [int] NULL,
    CONSTRAINT [PK_$(TableName)Out] PRIMARY KEY CLUSTERED
    [Partition_PK] ASC,
    [GUID_PK] ASC
    ) ON $(TableName)_fg2;
    GO
    -- 002 - Switch out partition in range 0-14
    --:SETVAR TableName "TestTable"
    ALTER TABLE dbo.$(TableName)
    SWITCH PARTITION 2 TO dbo.$(TableName)Out;
    -- 003 - Merge range 0 - 29
    :SETVAR TableName "TestTable"
    ALTER PARTITION FUNCTION $(TableName)_func()
    MERGE RANGE (14);
    -- Confirm table partitioning
    -- Original source of this query - http://lextonr.wordpress.com/tag/sys-destination_data_spaces/
    SELECT
    N'DatabaseName' = DB_NAME()
    , N'SchemaName' = s.name
    , N'TableName' = o.name
    , N'IndexName' = i.name
    , N'IndexType' = i.type_desc
    , N'PartitionScheme' = ps.name
    , N'DataSpaceName' = ds.name
    , N'DataSpaceType' = ds.type_desc
    , N'PartitionFunction' = pf.name
    , N'PartitionNumber' = dds.destination_id
    , N'BoundaryValue' = prv.value
    , N'RightBoundary' = pf.boundary_value_on_right
    , N'PartitionFileGroup' = ds2.name
    , N'RowsOfData' = p.[rows]
    FROM
    sys.objects AS o
    INNER JOIN sys.schemas AS s
    ON o.[schema_id] = s.[schema_id]
    INNER JOIN sys.partitions AS p
    ON o.[object_id] = p.[object_id]
    INNER JOIN sys.indexes AS i
    ON p.[object_id] = i.[object_id]
    AND p.index_id = i.index_id
    INNER JOIN sys.data_spaces AS ds
    ON i.data_space_id = ds.data_space_id
    INNER JOIN sys.partition_schemes AS ps
    ON ds.data_space_id = ps.data_space_id
    INNER JOIN sys.partition_functions AS pf
    ON ps.function_id = pf.function_id
    LEFT OUTER JOIN sys.partition_range_values AS prv
    ON pf.function_id = prv.function_id
    AND p.partition_number = prv.boundary_id
    LEFT OUTER JOIN sys.destination_data_spaces AS dds
    ON ps.data_space_id = dds.partition_scheme_id
    AND p.partition_number = dds.destination_id
    LEFT OUTER JOIN sys.data_spaces AS ds2
    ON dds.data_space_id = ds2.data_space_id
    ORDER BY
    DatabaseName
    ,SchemaName
    ,TableName
    ,IndexName
    ,PartitionNumber
    The table below shows the results of the ‘Confirm Table Partitioning’ query, before and after the MERGE.
    The data in the File and File Group to be dropped (File Group 2) has already been switched out; File Group 3 contains the data so no index rebuild is needed to move data and complete the MERGE.
    RANGE RIGHT would not be a problem in a ‘Sliding Window’ if the same file group is used for all partitions, when they are created and dropped it introduces a dependency on full index rebuilds. Larger tables are typically partitioned and a full index rebuild
    might be an expensive operation. I’m not sure how a RANGE RIGHT partitioning strategy could be implemented, with an ascending partitioning key, using multiple file groups without having to move data. Using a single file group (multiple files) for all partitions
    within a table would avoid physically moving data between file groups; no index rebuild would be necessary to complete a MERGE and system views would accurately reflect the physical location of data. 
    If a RANGE RIGHT partition function is used, the data is physically in the wrong file group after the MERGE assuming a typical ascending partitioning key, and the 'Data Spaces' system views might be misleading. Thanks to Manuj and Chris for a lot of help
    investigating this.
    NOTE 10/03/2014 - The solution
    The solution is so easy it's embarrassing, I was using the wrong boundary points for the MERGE (both RANGE LEFT & RANGE RIGHT) to get rid of historic data.
    -- Wrong Boundary Point Range Right
    --ALTER PARTITION FUNCTION $(TableName)_func()
    --MERGE RANGE (15);
    -- Wrong Boundary Point Range Left
    --ALTER PARTITION FUNCTION $(TableName)_func()
    --MERGE RANGE (14);
    -- Correct Boundary Pounts for MERGE
    ALTER PARTITION FUNCTION $(TableName)_func()
    MERGE RANGE (0); -- or -1 for RANGE LEFT
    The empty, switched out partition (on File Group 2) is then MERGED with the empty partition maintained at the start of the range and no data movement is necessary. I retract the suggestion that a problem exists with RANGE RIGHT Sliding Windows using multiple
    file groups and apologize :-)

    Hi Paul Brewer,
    Thanks for your post and glad to hear that the issue is resolved. It is kind of you post a reply to share your solution. That way, other community members could benefit from your sharing.
    Regards.
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Lead Time - Customizatiou0131n steps: Shipping point, route, and loading group

    It says about delivery scheduling that "The lead time considers the shipping point, route, and loading group". Can anybody explain me briefly the customization menu steps about this sentence?
    Thanks in advance.

    Please go through this document it will help you
    1) Maintain transportation planning point through Menu as follows:
    IMG Maintain&#61664; Logistic Execution &#61664; Definition &#61664;Enterprise Structure &#61664; transportation planning point.
    Create your own like BP01 enter the general data¡¯s required
    2) Defining modes of transport:
    &#61664;IMG Routes&#61664;Basic Transportation Functions &#61664; Transportation &#61664;Logistic Execution Define modes of transport&#61664;Define routes&#61664;
    Use the standard like:
    Shipping type: 01
    Description: Road
    3) Define shipping types :
    Basic Transportation&#61664; Transportation &#61664;Logistic Execution&#61664;IMG Define shipping types&#61664; Route &#61664;Functions
    Or through T-code: SM30
    PT: 01
    Description: Truck
    Mdtr: 01
    STPG: 0001
    4) Define transportation connecting point :
    Logistic&#61664;IMG Define&#61664;Routes &#61664; Basic Transportation Functions &#61664; Transportation &#61664;Execution Define transportation connection points&#61664;routes
    Copy the std point and make it FB01 as departure zone and FB02 as destination zone
    5) Define routes and stages.
    &#61664; Logistic Execution &#61664;IMG Define&#61664; Define routes &#61664;Routes &#61664; Basic Transportation Functions &#61664;Transportation routes and stages
    Use the standard route and copy it as FB0001 and enter the departure zone and destination as what we created.
    6) Defining Transportation Zone:
    Basic&#61664; Transportation &#61664;Logistic Execution &#61664;IMG Define transportation&#61664; Router determination &#61664; Route &#61664;Transportation Functions zones
    Use the standard transportation zone depending on the route so as for the example use 0000000003 - Southern Region
    7) Maintain country or transportation zone for shipping points
    &#61664; Logistic Execution &#61664;IMG &#61664; Route determination &#61664; Route &#61664; Basic Transportation Functions &#61664;Transportation Maintain country and transportation zone for shipping point
    Shipping point be ASH1 and route as 0000000003 which is assigned to country India
    8) Maintain route determination:
    Logistic Execution&#61664;IMG Route determination&#61664; Route &#61664; Basic Transportation Functions &#61664; Transportation &#61664; Maintain route determination&#61664;
    Country of departure: India
    Departure Zone: 0000000003
    Country of destination: India
    Destination Zone: 0000000003
    And enter the data’s needed like shipping condition 01, etc.
    Maintain Transportation relevance for shipping point, sales document type and delivery types.
    9) Now create shipping types
    &#61664;IMG Define shipment&#61664; Shipments &#61664; Transportation &#61664;Logistic Execution types
    Copy the standard 0001 and make it to AFIL
    10) Define and assign traffic zones through T-Code: T_76 (or) thru menu as follows
    &#61664;IMG Define&#61664; Basic functions &#61664; Shipments cost &#61664; Transportation &#61664;Logistic Execution and assign tariff zones
    Copy the standard and make it to FABTARIFF and assign our transportation planning point as ATP1, ATP2 assign tariff zone to every particular transportation planning point
    11) Define and assign pricing procedures
    Shipments&#61664; Transportation &#61664; Logistic Execution &#61664;IMG define and assign pricing&#61664; Price control &#61664; Pricing &#61664;cost procedures
    Copy the standard shipping cost pricing procedure SDFC00 to FBSC01
    12) Assign purchasing data:
    &#61664; Logistic Execution &#61664;IMG Assign purchasing data&#61664; Settlement &#61664; Shipments cost &#61664;Transportation
    13) Assign transportation planning point to shipping condition, purchase organization, purchasing group and to plants.
    14) Define Automatic G/L account determination:
    &#61664; Transportation &#61664; Logistics Execution &#61664;IMG Automatic G/L Account Determination&#61664; Settlement &#61664;Shipment Costs
    Or through T-code: OMWB
    Use G/L account 614000 as used in standard
    15) Automatic determination of CO assignments:
    &#61664;IMG Automatic&#61664; Settlement &#61664; Shipment Costs &#61664; Transportation &#61664;Logistics Execution Determination of CO Assignments
    Or T-code: OKB9
    16) Maintain shipping point address.
    &#61664; Logistics Execution &#61664; Definition &#61664;Enterprise Structure Define, copy, delete, check shipping point
    17) Creating G/L account centrally.
    Financial Accounting&#61664;Accounting &#61664; Master Records &#61664; General Ledger &#61664;
    &#61664;Individual Processing Centrally
    Or through T-Code: FS00
    18) Enter the G/L account 614000 to the company code AFIL
    19) Create transportation condition records:
    &#61664; Shipment cost &#61664; Transportation &#61664; Master data &#61664;Logistic Execution create&#61664;Prices
    Or thru T-code: TK11
    Enter condition type FB00 that is freight condition type, which is used as standard in the shipping cost pricing procedure.
    Enter the price and save it
    Process:
    1. The system determines the pricing procedure for the shipment cost item-using transportation planning point, service agent and shipment cost item category
    2. Calculation Basis
    3. Condition Record
    4. Access sequence.
    For the sales order, create a delivery using VL01N and pick the goods. Now create a shipment using VT01N and assign the delivery to the shipment document. Make sure to enter the carrier and other details before performing the transportation planning and execution. Save the shipment document and PGI for the delivery would be done automatically.
    A shipment represents a physical goods movement between two or more locations.
    For the system to create a shipment, shipping-relevant deliveries must exist.
    You use Transportation Planning to group pending deliveries into shipments and then to perform the shipments.
    The outbound delivery documents contain routes that the system calculates in route determination, and a transportation planning date for the delivery.
    Dynamic transportation planning is responsible for grouping together outbound deliveries into shipments that have at least the same route, visit plan type and the same transportation planning date.
    The following restrictions can be taken into account for a shipment in Dynamic Transportation Planning:
    • Weight
    • Volume
    • Variable capacity
    • Number of stops
    • Maximum duration
    • Loading units
    Create Shipment Document
    Use
    After the status of the delivery is open for transportation planning, it is possible to create a shipment document to schedule the transportation.
    Transportation&#61664; Create &#61664;Single document Logistics&#61614; Logistic Execution &#61664; Transportation&#61664;Planning
    T.code: VT01n
    Enter your transportation planning point and shipment type. Press enter.
    Creating the shipment cost document
    Use
    After the deliveries are scheduled by shipment and the shipment is executed, now it is time to maintain and close the shipment document in order to record all relative information and ready for the shipment cost settlement.
    &#61664;Single document Logistics&#61614; Logistic Execution &#61664; Transportation&#61664; Shipment cost &#61664;Create
    T.code: VI01.
    After creation of shipment cost doc. You settle the Shipment cost.
    Go to VI02, enter shipment cost no.and press enter.
    Then go to Overview -- Fright cost items, Select the line item and double click it.
    Go as main menu: Item --> Acc. assignment, you enter G/L acc. and cost center to which the transportation cost will be posted.
    On the Standard change, details – Freight cost items screen, then check the check box “Transfer”, which means you will transfer the relative transportation cost to FI/CO, in this case, the system should determined which Purchase Order will used to settle the cost, so after you check the “Transfer” check box.
    Press save button on the current screen.
    Shipment cost settlement and documents:
    Use
    After the shipment cost settled in shipment cost document, it has the following affects:
    It creates a service entry sheet that refers to the purchase order determined.
    It posts acceptance of services performed.
    It creates an accounting document for the accruals.
    Further subsequent documents are created in FI.
    Create Delivery:
    T.code: Vl01n
    Create TO:
    T.code: LT03
    Create Shipment Document
    Use
    After the status of the delivery is open for transportation planning, it is possible to create a shipment document to schedule the transportation.
    &#61664; Transportation Planning &#61664; Create &#61664;Single document Logistics&#61614; Logistic Execution &#61664;Transportation
    T.code: VT01n
    Enter your transportation planning point and shipment type. Press enter.
    Creating the shipment cost document
    Use
    After the deliveries are scheduled by shipment and the shipment is executed, now it is time to maintain and close the shipment document in order to record all relative information and ready for the shipment cost settlement.
    &#61664;Single document Logistics&#61614; Logistic Execution &#61664; Transportation&#61664; Shipment cost &#61664;Create
    T.code: VI01.
    After creation of shipment cost doc. You settle the Shipment cost.
    Go to VI02, enter shipment cost no.and press enter.
    Then go to Overview -- Fright cost items, Select the line item and double click it.
    Go as main menu: Item --> Acc. assignment, you enter G/L acc. and cost center to which the transportation cost will be posted.
    On the Standard change, details – Freight cost items screen, then check the check box “Transfer”, which means you will transfer the relative transportation cost to FI/CO, in this case, the system should determined which Purchase Order will used to settle the cost, so after you check the “Transfer” check box.
    Press save button on the current screen.
    Shipment cost settlement and documents:
    Use
    After the shipment cost settled in shipment cost document, it has the following affects:
    It creates a service entry sheet that refers to the purchase order determined.
    It posts acceptance of services performed.
    It creates an accounting document for the accruals.
    Further subsequent documents are created in FI.
    Generally process is-
    1) VL01N
    2) VT01
    3) VL02N - PGI
    4) VI01
    5) ML81N – Service Entry Sheet
    6) MIRO – Invoice Verification
    Shipping docs (VT01) are created before PGI...
    You cannot carry out transportation planning or creation of shipment after the goods issue has been executed. Once goods issue is done it means that the material has left your premises so there is no need for a transportation planning after it.
    In case you need carry out the shipment documents at the end of the day you could just create the delivery documents during the day and at end of the day create the shipment documents and the goods issue
    This link is also usseful:
    http://help.sap.com/printdocu/core/Print46c/en/data/pdf/LETRA/SDTRA.pdf
    Hope this helps you.
    Regards
    nitin

  • Which tool do you use for load testing?

    Which tool do you use for load testing for J2EE application?

    hi,
    i use jmeter for without issues..
    first make sure that the system from where you are doing load testing is in the same lan(without firewall) to that of server. Here you are testing the capacity of server and not the network bandwith.
    So assuming very good network connection what is the servers performance level...
    If u have to decide which tool is decent enough to load testing,
    i would suggest go for jmeter. its reliable and i have used it without problems.
    Here is my quick tutorial for jemeter which i made for my teams reference.
    REFERENCE
         For Getting Started refer to any tutorial for Jmeter
         You can refer to introduction on
              http://jakarta.apache.org/jmeter/usermanual/
    RUNNING LOAD TEST
         Extract jakarta-jmeter-2.3RC4.zip in c:
         install jdk-1_5_0_12-windows-i586-p.exe
         Run C:\jakarta-jmeter-2.3RC4\bin\jmeter.bat
         In Test Plan add Thread Group
         Right Click Thread Group -> Add->Sampler->HTTP Request
         Select HTTP Request added fill details
              Serverip:www.xyz.com
              Port:80
              Path:/xyz-webapp/index.jsp
              Use KeepAlive: Checked
              Retrive All Embeded images from HTML: Checked(For geting images inside html)
         Right Click Thread Group -> Add->Listner->Graph Results
         Click on Thread Group Fill details
              No of Threads:3(This means no of simultaneous users)
              Ramp Up Period:1(3 Request will be made in 1 second period)
              Loop Count:5(Above process should be repeated how many times)
         Save TestPlan
         Run -> Clear All
         Run -> Start
              Click on Graph Results for Output
         Try adding different listener in Thread Group
              Thread Group -> Add->Listner->Summary Report
         Listeners coud be added at thread group level or HTTP Request Level
    RECORDING A TEST.
         Add a HTTP proxy server in workbench by right clicking on it and selecting Add->Non test Element->HTTP Proxy Server.
         In Test Plan Add Thread Group
         In HTTP proxy server select Target Controller as Test Plan > Thread Group.
         Go to Internet Explorer and go to internet settings and set proxy server as localhost and the port
              which is configured in Jmeter proxy server.Save Settings.
         In Jeter proxy server start by Clicking on Start button in the HTTP Proxy Server Flap.
         In browser navigate the application which you want to test.
         The nodes will be added in the Thread Group.
         Once Navigation is done stop the proxy server and save the test plan from file menu.
    LOGGING INTO APPLICATION.
         When we have to login into system and then navigate the system
              we need a cookie to be tracked after login. This could be achieved
              simply by Add -> Config Element -> HTTP Cooie Manager.
         After adding the HTTP Cooie Manager simply run the test.
    LOOPING COUNTER.
         When a request is made multiple parameter is sent to server.
         Parameters are sent from the HTTP Request node added in the Thread Group
              in "Send Parameter with the request" table.
         If we run the loop count 10 times the parameters will be sent same in every request.
         If you have to send parameters like user1,user2,user3,user4....
              then add -> Pre processor -> Counter.
         Put start as 1 and and Maximum as 10 or whatever required.
         Put Referene name as iCtr or whatever.
         In the "Send Parameter with the request" in value column put user${iCtr}

  • Crystal Reports XI R2 Server problem when trying to load groups from LDAP

    Hi,
    I'm using Crystal Reports XI R2 SP4 on RedHat Linux. My users are being authorized through LDAP (OID). My problem is that after removing all groups (right now no groups/users are visible in CMC) in CMC Authentication tool and then trying to add them again I'm receiving following error:
    "LDAP: Failed to commit objects to server : Duplicate object name in the same folder."
    I've checked connectivity to LDAP server and it's ok. Could it be connected to CMS problems/corruption? Any help/suggestions would be appreciated.
    Regards,
    Marcin

    The only personal folders visible from under Admin panel are for users Administrator and Guest. I've tried to query CMS tables but these store entries not in plain text so I cannot identify entries connected to personal folders. I have also tried to do this via Query Repository tool but all returned entries were ok.
    I have also noticed that when trying to load group from LDAP after it returns error "LDAP: Failed to commit objects to server : Duplicate object name in the same folder." entry for this group is created in CMS table. Also this group is then visible in admin Groups panel. No users are created with the group and when I'm trying to list users belonging to added group it returns "internal secLDAP security plug-in error" but without any other suggestion what went wrong.

  • Shipping point 01 loading group plant storage location not defined.

    Hi All,
    I have a unique problem where iam getting the above error when sales order is created using incomplete log.
    I have checked all the settings related to Shipping point
    in OVL2  Shipping point are assigned,
    in OVL3 picking location is also assigned.
    I have been searching entire SDN and help.sap.com for the above error for last three days and post a new thread.
    Please help, points rewards.
    Regards,
    Vinod.

    Dear Vinod
    First of all I am really happy that you have taken the trouble to find out a solution to your problem and since you were not able to get it, you had posted a new thread.  Of late, this is very rare people searching the forum b4 posting. 
    Coming to your problem, go to customer master Sales Area tab and see in Shipping tab, what Shipping Condition is maintained.  Ensure that the same is maintained in OVL2.
    Similarly, the same for Loading Group.  Normally, by oversight, people commit mistakes by maintaining differently which is not assigned in masters and this would have missed your attention.  For example, you would have assigned 01 in customer master and in OVL2, you would have maintained 03.
    Also, please ensure that in OVXC, you have assigned shipping point to your plant.
    thanks
    G. Lakshmipathi

  • Load Testing Discoverer Plus 11g Web Using HP Load Runner 11

    Hi,
    I am doing load test for "Oracle Business Intelligence Discoverer Plus 10g" using HP Load Runner 11. Could you tell me which is the best protocol to record script?
    I am looking for a protocol which I can parametrize after recording.
    These are recommended protocols from Load Runner when I run Protocol Analyzer
    AJAX(Click and Support) : Seems to work fine but Can't parametraize Query after recording
    Java Over HTTP
    Web (HTTP/HTML)
    There are other protocols too.
    Could you provide any pointers or best practices. Thanks a lot
    Edited by: 936062 on May 22, 2012 2:56 PM

    Hi,
    You can refer this forum
    [http://groups.google.com/group/LR-LoadRunner/browse_thread/thread/f1fb1c6430e39f94/565baac51a6d35c0]
    Thanks and Regards,
    Dhaya.G.

  • How do I use Panorama / Tab Groups with the keyboard?

    Yes I know that CTRL-SHIFT-E opens the panorama window, and then I can use the mouse to organize my tabs into groups.
    But how do I do this with the keyboard?
    I've [http://lifehacker.com/#!5719596/switch-between-tab-groups-in-firefox-with-a-keyboard-shortcut read] [http://ubuntuforums.org/showthread.php?t=1705714 that] Ctrl-`should move me through tab groups. That doesn't work for me on my Danish keyboard. (Where the ' and ` chars are weird. But is how they are on a valid standard Danish keyboard) I've tried changing the keyboard to USA and then moving through tab groups works fine.
    In short: Pretend I don't have a mouse. How do I use Panorama / Tab Groups?

    Sorry. These are both known bugs:
    [https://bugzilla.mozilla.org/show_bug.cgi?id=587010 Bug 587010] - Add keyboard UI for manipulated groups and tabs in Tab Candy
    and
    [https://bugzilla.mozilla.org/show_bug.cgi?id=626594 Bug 626594] - shortcut for switching tab groups is badly accessible on many non-US keyboard layouts

  • How to load a default value in to a column when using sql loader

    Im trying to load from a flat file using sql loader.
    for 1 column i need to update using a default value
    how to go about this?

    Hi!
    try this code --
    LOAD DATA
       INFILE 'sample.dat'
       REPLACE
       INTO TABLE emp
       empno   POSITION(01:04) INTEGER EXTERNAL NULLIF empno=BLANKS,
       ename   POSITION(06:15)  CHAR,
       job         POSITION(17:25)  CHAR,
       mgr       POSITION(27:30)  INTEGER EXTERNAL NULLIF mgr=BLANKS,
       sal        POSITION(32:39)  DECIMAL EXTERNAL NULLIF sal=BLANKS,
       comm   POSITION(41:48)  DECIMAL EXTERNAL DEFAULTIF comm = 100,
       deptno  POSITION(50:51)  INTEGER EXTERNAL NULLIF deptno=BLANKS,
       hiredate POSITION(52:62) CONSTANT SYSDATE
      )-hope this will solve ur purpose.
    Regards.
    Satyaki De.

Maybe you are looking for

  • SharePoint 2013 Multi-tenant Feature Packs

    I have not seen any information in regards to multi-tenancy feature packs for SharePoint 2013. Currently it seems, with the new end user licencing, when you install the Enterprise bits you can license a user for either Standard or Enterprise but not

  • Can't send mail using webmail on Mac OS X Server 10.4

    Could someone help me figure out why I can't send email via webmail on Mac Mac OS X Server 10.4. I tried in and out of my LAN but it will not work. I am pretty sure that it's no firewall issue because it won't even work inside the LAN. Another thing

  • In need of help, creating a semi- simple program

    I'm creating a program that is suppose to fit such criteria, Read a file of real numbers (doubles), find the smallest and largest, print the numbers read (4 per line), and print the following statistics for the input data: · Sum · Smallest · Largest

  • Want to download itunes 11.1?

    help me download itunes 11.1 version please

  • 5D Markiii and a green color cast

    Does anyone else experience a green cast/tint to their images while shooting in JPG with a 5DMkiii? I shoot using AWB and have even adjusted the color on the quick screen histogram to try and compensate....but skintones consistantly look green. My mo