GZIP and External command.

Hello,
My problem might seem simple but I have tried for a couple of days now to make researches on that matter and it still doesn't work, no matter what.
I had to make, with SM69, two external commands. One was using the command "mv" which is working very well and the other must be a "gzip" with a parameter "-c" and, in a module function, with the function "SXPG_COMMAND_EXECUTE_LONG", I have to send the path and the file name.
So since I am not able to make it work in the module function I'm creating, I am working to make it work on SM69. So here what I have entered so far...
Operating System Command: gzip
Parameters for operating sytem command: -c
Additional parameters: path/file.txt > newpath/file.txt.gz
Though, the only thing I am receiving is no change at all and I'm receiving  u2039  È ®J as a message. The Exit code is at 1 and the exit status is E.
I know it means there is something wrong but all my researches lead me to do it that way so I'm not sure where I am going wrong and how to make it work appropriately.
Can you please help me?
I am open to questions if you have any.
Thank you

Hi, kindly check the following  example
Command name : ZTEST
Operating system command : cmd /c
v_dir_input = 'cmd /c c:\winzip\wzzip.exe -password  g:\test1.zip g:\test.pdf'.
CALL FUNCTION 'SXPG_COMMAND_EXECUTE'
  EXPORTING
    commandname                   = c_extcom
    additional_parameters         = v_dir_input
    operatingsystem               = c_oper
  TABLES
    exec_protocol                 = t_result
  EXCEPTIONS
    no_permission                 = 1
    command_not_found             = 2
    parameters_too_long           = 3
    security_risk                 = 4
    wrong_check_call_interface    = 5
    program_start_error           = 6
    program_termination_error     = 7
    x_error                       = 8
    parameter_expected            = 9
    too_many_parameters           = 10
    illegal_command               = 11
    wrong_asynchronous_parameters = 12
    cant_enq_tbtco_entry          = 13
    jobcount_generation_error     = 14
    OTHERS                        = 15.

Similar Messages

  • Problem with data table and internal and external command button

    thanks for your time,
    i have this problem
    <t:dataTable binding="#{manejadorAsociarEvaluadoresProyecto.tablaPosiblesEvaluadores}"
    cellpadding = "6"
    cellspacing = "3"
    value = "#{manejadorAsociarEvaluadoresProyecto.listaPosiblesEvaluadores}"
    var = "posibleEvaluador"
    preserveDataModel="false">
    <h:inputTextarea value="#{posibleEvaluador.correoActualCuerpo}"
    cols="50"
    rows="50"
    rendered="#{posibleEvaluador.detalle}" />
    <h:commandLink immediate="false" action="#{posibleEvaluador.editar}" value="Editar Correo"/>
    </t:dataTable>
    <h:panelGroup id="panelBotonEvaluadores" rendered="#{!manejadorAsociarEvaluadoresProyecto.b_mostrarEvaluadores && manejadorAsociarEvaluadoresProyecto.tieneProyectos}">
    <h:commandButton action="#{manejadorAsociarEvaluadoresProyecto.mostrarEvaluadores}"
    value="Asociar evaluadores"/>
    </h:panelGroup>
    this is the method of posibleEvaluador
    public String editar()
    setDetalle(!getDetalle());
    return "";
    all run perfect, until y click in "asociar evaluadores" button, after of this the commandLink "editar Correo" not run, this commanlLink is for not show the inputTextArea, i don't know what happend.
    thanks for your help

    Hello,
    I have replicated and confirmed the problem with LabVIEW 2010, Report Generation Toolkit 2010 and Word 2010.  I have filed this to R&D under CAR ID #257414.  As a workaround you can manually set each cell using Word Edit Cell.vi.  I have attached a simplified representation of this unexpected behavior as well as the workaround.  Hopefully this helps everyone out!
    David_L | Certified LabVIEW Architect
    LabVIEW Tools Network | LabVIEW Tools Network Developer Center
    Attachments:
    Word 2010 Table CAR.vi ‏17 KB
    Word 2010 Table FIX.vi ‏22 KB

  • Executing external command and program in background

    Hi,
    how to execute external command and external program in background.
    please provide me some tips on the same.
    while creating external command, what is the significance of "operating system command" and "parameters for operating system command".
    Thanks,
    Dinesh.

    Hi,
    You can define external commands using transaction code SM69.
    External commands are commands which would be executed at your operating 
    system with or without parameters.
    Operating system commands are those which you execute at OS level like
    mkdir, cat, rm, cp, ls
    Parameters are options that you add to os commands to get more functionalities.
    like, < ls -l > : will give you a long listing of files where " l " is the parameter
            < ls -lt > : will give you a long listing with last changed file at top where " lt " is the parameter.
    You can mention these in SM69 when you define the external commands.
    Now, in order to execute these in your background job. In the steps click on "external command" and give your command name in SM36 while creating your job.
    " Points for solution ".
    Thanks and Regards,
    Sandeep.

  • A better way to execute a series of external commands?

    Hi,
    Consider the code snippet...
    Process proc1 = rt.exec("Some external command");
    Vector vecOutput = outputGobbler.getOutput() ;
    String[] strAlterCmds = new String[vecOutput.size()];
    for ( int k = 0 ; k < vecOutput.size() ; k++ )
    strAlterCmds[k] = new String();
    strAlterCmds[k] = "cmd.exe /c blah.... " +(String) vecOutput.elementAt( k )+ " ;
    Process proc2 = rt.exec(strAlterCmds[k], null);
    StreamGobbler errorG = new
    StreamGobbler( proc2.getErrorStream(), "Error") ;
    StreamGobbler outputG = new
    StreamGobbler( proc2.getInputStream(), "Output") ;
    errorG.start() ;
    outputG.start() ;
    int exitVal1 = proc2.waitFor();
    The second part of the execution is taking AGES and I am not sure forking a new process for runtime execution is the best alternative here.
    I would be glad if someone can point me to a smarter solution.
    In short: I intend to execute a series of commands using RTE depending on the values I get in the loop.
    Thanks.

    Jared,
    Thank you for responding to my posted message. Rendezvous is a new concept to me, maybe
    it's the solution to my problem. I have been trying to read the on line manual and
    example VIs from National Instruments website. I still have a hard time to understand
    the concept.
    One of the example I saw is using rendezvous to run some sub VIs. But in my case, I have
    only one VI that is a while loop. Inside the while loop, there are a few tasks running
    simultaneously. I don't know whether it will serve my purpose.
    Guangde Wang
    Jared O'Mara wrote:
    > Guangde,
    >
    > Referring to your 2nd method, use rendezvous (under advanced>>synchronize
    > on function palette) to synchronize two processes. There are good examples
    > that come with labview. basically, you cre
    ate a rendezvous, giving it a
    > size. Using the wait on rendezvous vi, a function will not continue until
    > all rendezvous have reached that point. Using this method, you can synchronize
    > your 2 while loops.
    >
    > Jared
    >
    > Guangde Wang wrote:
    > >I tried two ways to control the tempo of my program.>>One is to use the
    > While Loop's counter and a wait. The drawback of this>method is the cycle
    > length is dependent on the measuring load. So if the>program runs for days,
    > it will be significent off the real time.>>The other way is to use the difference
    > of the Tick Count. It provides>accurate timing but the problem is the sychronization
    > of the clock cycle>and the While Loop cycle. I can try to put a little bit
    > wait but still>can not sychronize them very well. So after a while, there
    > will be two>measures get together.>>I don't know whether there are some better
    > ways to control the program>or whether we have some ways to improve either
    > or both of the above
    two>methods to make them work better. Please let me
    > know if you have any>suggestion.>>Thank you in advance,>>Guangde Wang>

  • How do you highlight multiple documents rapidly, rather than having to click on the document and on "command" one at a time?

    How do you highlight multiple documents rapidly, rather than having to click on the document and on "command" one at a time? Right now, when I'm trying to highlight documents to export, I have to click on each document individually. Is there a way to highlight one document while scrolling down the list until all desired documents are highlighted? I would like to rapidly be able to do this to save time when I am exporting or saving multiple files from my documents to an external drive, CD-Rom or cloud storage.

    Click on the first item
    Hold the Shift key
    Click on the Last Item.
    All the items in between will be selected.

  • SharePoint 2013 - Office Web Apps - Internal and External Use

    I have successfully installed SharePoint 2013 and Office Web Apps on Azure VMs inside an Azure Virtual Network (IaaS model). Everyting is working well. However, my testing has shown that external users and internal users can't use Office Web Apps at the
    same time.
    Office Web Apps, installed on its own vm, accomodates an external and internal URL quite well. However, SharePoint 2013 appears to only allow one setting for WOPI Zone, either internal or external but not both. I've set the WOPI zone to Internal-HTTPS (Set-SPWOPIZone
    –Zone “internal-https”). OWA works just fine if accessed from inside the Azure Virtual Network. However, if I try to access from outside the Virtual Network, from the Internet, Office Web Apps fails. The exact oppisite is also true. I can set WOPI Zone to
    External-HTTPS and accessing from the Internet works fine, but accessing inside the Virtual Network fails.
    Am I missing something? I, obviously, want Office Webs Apps to function properly for both internal and external users simultaneously.
    I appreciate any help anyone can provide here.
    Glenn

    Hi Glenn,
    To have both the use of Internet and Internal available to your end-users, you first need to configure AAM setting. Open Central Administration > Application Management > Configure alternate access mappings. Let's say there is an existing web application
    named http://sharepoint and my end-users from local network are able to access it using the URL http://sharepoint (root site collection). Here you need to add the Internet URL by select the web application and click Edit Public URLs. Add the Internet domain
    to the web application, e.g http://sharepoint.abc.com. You don't necessarily have to edit binding setting in IIS. Before continuing next steps, make sure you are able to access http://sharepoint.abc.com from the Internet while being able to access http://sharepoint
    from local network (aka Internal).
    On the machine where Office Web App (OWA) Server 2013 is installed, open PowerShell to add OWA module and use the following command to re-create a new OWA server farm if you've completed configuring it previously.
    New-OfficeWebAppsFarm -InternalUrl "http://owa" -ExternalUrl "http://owa.abc.com" -EditingEnabled.
    In this case, I'm not using SSL certificate to encrypt data over the Internet. You can use Internet-public IP of the OWA server like -ExternalUrl "http://198.xxx.xxx.xx". Add CertifcateName parameter if you want to use whether CA-issued certificate
    or self-signed certificate.
    On your SharePoint machine, you need to re-bind all WFE machines to WAC farm using the cmdlet New-SPWOPIBinding. Next, you need to set the WOPI zone for both internal and external.
    Set-SPWOPIZone -zone "external-http"
    Note: I'm not all using certificate in my guidance. But the steps to have it configured is just to add more parameter. 
    I've recently successfully deployed OWA multi-server farm for both internal and internet uses for two big clients. In real-world scenario, ideally OWA should be published through firewall (Forefront UAG, TMG, F5...etc). Please let me know if you still have
    issues after following my steps. My email: [email protected]
    Regards,
    -T.s
    Thuan Soldier
    A 23-year-old man loving Microsoft technologies and making crazy ideas on business journey.
    SharePoint Vietnam |
    Blog | Twitter

  • Using a Logical File (FILE) in an External Command (SM69)

    I have an external command defined in SM69 that runs a script on the application server. The command is called from within an ABAP program with SXPG_CALL_SYSTEM. It currently works as it should.
    The issue is that the external command is defined as:
    sh /absolute/path/to/script/file
    I have been told to instead set up a logical file in FILE so that the command would be:
    sh <MY_SCRIPT_LOGICAL_NAME>
    I've gone into FILE and defined MY_SCRIPT_LOGICAL_NAME as:
    Logical File = MY_SCRIPT_LOGICAL_NAME
    Name = Descriptive Text
    Physical File = <SYSID>actualname.scr
    Data Format = ASC
    App Area = HR
    Logical Path = SCRIPT_PATH
    where SCRIPT_PATH is also defined in FILE with the physical path of:
    /iface/<SYSID>/outboundhr/scripts/<FILENAME>
    So when I try executing the command with SM49 as defined as:
    sh <MY_SCRIPT_LOGICAL_NAME>
    I just get Exception raised: OTHERS
    If I define the command as:
    ls <MY_SCRIPT_LOGICAL_NAME>
    I get the same thing, Exception raised: OTHERS
    I'm guessing that the logical file name isn't getting translated. Does anyone know if it should in this instance? Is anyone using a logical file as part of an external command?
    Thanks much.

    I think you must use FM FILE_GET_NAME to translate the logical file name to a physical one before executing the command.
    Like:
    DATA: physical_file_name(60),
          my_script_logical_name(60) VALUE 'CLASSIFICATION',
          format(3).
    CALL FUNCTION 'FILE_GET_NAME'
      EXPORTING
        logical_filename = my_script_logical_name
      IMPORTING
        file_format      = format
        file_name        = physical_file_name
      EXCEPTIONS
        file_not_found   = 1
        OTHERS           = 2.
    Rob
    Message was edited by: Rob Burbank

  • External command is not working in File Receiver Adapter

    Hi,
       I have tried to copy the file which is mentioned in my File Adapter (Receiver) to 3 different folders thru OS command. The purpose is to save time. I have created one CC for Source(File) and one CC for target(File). After the target file is created, I want to copy this file in 3 folders. For this I have written one batch file in my system.
    File Name: copy1.bat
    @ECHO OFF
    COPY %1 C:\TEST1\arch_t1
    COPY %1 C:\TEST2\arch_t2 
    COPY %1 C:\TEST3\arch_t3
    DEL %1
    The command which I entered in the Receiver CC (File Adatper is)
    Run Operating System Command After Message Processing
    Command Line: C:\FILES\copy1.bat %F %f
    Timeout(secs): 60
    Terminate Program After Timeout - Check box is selected.
    But, this external command is not working. (because the target files are not created in those directories C:\TEST1, C:\TEST2, C:\TEST3)
       I did this scenario by refering the Michael'w weblog:
    /people/michal.krawczyk2/blog/2007/02/08/xipi-command-line-sample-functions
       In this blog, he has also given how to troubleshoot external command. I tried this also. It is not working.
        I use Windows XP OS. XI 3.0 SP 18.
       Friends, Could you kindly tell why this external command is not working? What could be the problem?
    Thanks in advance.
    Kind Regards,
    Jeg P.

    Hi, Jeg.
    As the note says, please check the java Runtime.exec behavior
    with the proper user.
    If in doubt, check if the command can be successfully run through
    the Java Runtime.exec(...) API. Also note that the command is run
    as user "<sid>adm" (Unix) / "SAPService<SID>" (Windows).
    So, Could you check the following(very dirty one!) Java in your
    XI server and check if it works? For simulating %F, create a test
    file C:\FILES\test.txt whatever the content is.
    import java.util.*;
    import java.io.*;
    public class ExecJava {
      public static void main(String args[]) {
        try {           
          Runtime rt = Runtime.getRuntime();
          Process proc = rt.exec("C:/FILES/copy1.bat C:
    FILES
    test.txt");
        } catch (Exception e) {
          e.printStackTrace();
    Good Luck.
    Sejoon

  • What are the commands using in SAP ECC 6.0 ehp on SYBASE Database? Backup and restore commands?

    Hi All,
    What are the commands using in SAP ECC 6.0 ehp on SYBASE Database? Backup and restore commands?

    Hi Jayachander,
    Have a look at these sap notes:
    For taking backup: Schedule from DB13 and get the exact command from Logs
    1841993 - SYB: How to schedule backups in DBA Cockpit
    1887068 - SYB: Using external backup and restore with SAP Sybase ASE
    How to restore DB
    1611715 - SYB: How to restore a Sybase ASE database server (Windows)
    Divyanshu

  • Change status(Active/Inactive) of file adapter by the external command

    Hi, all.
      Is it possible to change status(Active/Inactive) of file adapter by the external command?
      Let's say, like
      "$ switch_file_adapter_status.sh <File Adapter name> <active|inactive>"
    or something like that.
      Actually, we have the following requirement for the file adapter.
      For example, we have many "file adapter to R/3" scenarios and we only want to allow data transfer between 9:00 a.m. to 9:00 p.m.(R/3 service time).
      In order to achieve this requirement, one option would be to switch file adapter status between active and inactive by external command line.
    And this command will be controlled by the job scheduler(like Tivoli).
      Does XI(3.0 or above) have this kind of feature?
      Best Regards.

    Hi,
    I don't think that this is possible.
    But one Solution for your problem could be a job on xi who execute the Queue.
    So that Adapter work the hole time and send the Message to XI Queue which process them not automatically.
    And the XI Queue you can trigger with a Job like normal R/3 Jobs.
    Hope that helps,
    Regards,
    Robin

  • Imac and external hdd

    im trying to boot from external hdd and efi bios does not even see the external hard drive!!!!
    i work at the educational facility and i just got 40 imacs that have to be cloned!!!
    doing every one of those by hand would be a nightmare.
    does apple actually provide any tools except netboot and netinstall that would help me with the quest?
    any tricks i have to do on the mack to recognize external drive as a boot drive?

    A good source for how to do netinstalls is www.bombich.com.
    You fail to tell us anything about this external drive, how it's formatted, what's installed on it, etc. The general rundown is:
    You need to partition the external drive using the GUID partition scheme. If the external drive has never been formatted before then format the drive Mac OS Extended (Journaled.) This should be done using the OS X Installer Disk One that came with the computers. If the computers are all the exact same model then you only need use one of the installer discs. However, if you have different hardware, then each different model has a different build of OS X. You will need two different external drives or you will have to update the system on the external drive to one that will be common for all the models you have (this would be 10.4.8 (Intel).
    Here's the procedure:
    Extended Hard Drive Preparation
    1. Boot from your OS X Installer Disk. After the installer loads select your language and click on the Continue button. Then select Disk Utility from the Installer menu (Utilities menu for Tiger.)
    2. After DU loads select your external drive (this is the entry with the mfgr.'s ID and size) from the left side list. Click on the Partition tab in the DU main window.
    3. Set the number of partitions from the dropdown menu (use 1 partition unless you wish to make more.) Click on the Option button and select the GUID partition map and click on OK. Set the format type to Mac OS Extended (Journaled.) Click on the Partition button and wait until the volume(s) mount on the Desktop.
    4. Select the volume you just created (this is the sub-entry under the drive entry) from the left side list. Click on the Erase tab in the DU main window.
    5. Set the format type to Mac OS Extended (Journaled, if supported.) Click on the Options button, check the button for Zero Data and click on OK to return to the Erase window.
    6. Click on the Erase button. The format process will take 30 minutes to an hour or more depending upon the drive size.
    After formatting is completed quit DU and return to the installer. Now quit the installer. The computer will reboot. After reboot clone the system on the iMac to the external drive:
    How to Clone Using Restore Option of Disk Utility
    1. Open Disk Utility from the Utilities folder.
    2. Select the backup or destination volume from the left side list.
    3. Click on the Erase tab in the DU main window. Set the format type to Mac OS Extended (journaled, if available) and click on the Erase button. This step can be skipped if the destination has already been freshly erased.
    4. Click on the Restore tab in the DU main window.
    5. Select the backup or destination volume from the left side list and drag it to the Destination entry field.
    6. Select the startup or source volume from the left side list and drag it to the Source entry field.
    7. Double-check you got it right, then click on the Restore button.
    8. Select the destination drive on the Desktop and press COMMAND-I to open the Get Info window. At the bottom in the Ownership and Permissions section be sure the box labeled "Ignore Permissions on this Volume" is unchecked. Verify the settings for Ownership and Permissions as follows: Owner=system with read/write; Group=admin with read/write; Other with read-only. If they are not correct then reset them.
    For added precaution you can boot into safe mode before doing the clone.
    Why reward points?(Quoted from Discussions Terms of Use.)
    The reward system helps to increase community participation. When a community member gives you (or another member) a reward for providing helpful advice or a solution to their question, your accumulated points will increase your status level within the community.
    Members may reward you with 5 points if they deem that your reply is helpful and 10 points if you post a solution to their issue. Likewise, when you mark a reply as Helpful or Solved in your own created topic, you will be awarding the respondent with the same point values.

  • External commands fail in Background Jobs on Application server

    Hi,
    We are having problems with jobs that include External Commands. The jobs run successfully if on the Central Instance. But if they run on either of the two Application Server's it fails with 'Unable to connect to Target Host' error message.
    I can run the External Commands from SM49 on both AS's. But as soon as its run in background it fails.
    I have read the rfcexec.sec file needs to be in the .\work directory on the application so I have this in place.
    This is what gets written to dev_w0 when the failure occurs :
    M  *****************************************************************************
    M  *
                                                                                    M  *  LOCATION    SAP-Gateway on host <APP_SERVER> / sapgw75
    M  *  ERROR       timeout during allocate
    M  *
                                                                                    M  *  TIME        Fri May 30 11:20:34 2008
    M  *  RELEASE     640
    M  *  COMPONENT   SAP-Gateway
    M  *  VERSION     2
    M  *  RC          242
    M  *  MODULE      gwr3cpic.c
    M  *  LINE        1795
    M  *  DETAIL      no connect of TP sapxpg from host <CENTRAL_INSTANCE.domain>
    M  *              after 22 sec
    M  *  COUNTER     483
    M  *
                                                                                    M  *****************************************************************************
    M
                                                                                    A  RFC 1446  CONVID 48742194
    A   * CMRC=27 DATA=0 STATUS=0 SAPRC=242 ThSAPECMINIT
    A  RFC> ABAP Programm: SAPLCRFC (Transaction: )
    A  RFC> User: <USERNAME> (Client: 300)
    A  RFC> Destination: %_TCPIP_%1 (handle: 2, , )
    A  RFC SERVER> RFC Server Session (handle: 1, 48741194, {C4A35AF0-2F4C-4E36-98C3-9999A6FE99D7})
    A  RFC SERVER> Caller host: iblonc4306_P75_75
    A  RFC SERVER> Caller transaction code:  (Caller Program: RSBTCXPG)
    A  RFC SERVER> Called function module: SXPG_STEP_COMMAND_START                                
    I'd really appreciate any suggestions on how to fix this issue??
    Thanks,
    Kye

    Hi Siddhesh,
    Thank you for your prompt response. I'm confident all of the prerequisites are set correctly.
    I have enabled the trace on SAPXPG but its only writing to the log dev_cp and dev_xpg when the command is executed successfully.
    For example, if I use SM49 to run the External command from one of the Application Servers dev_cp & dev_xpg  are updated and the command completes successfully. However if I run the job (that has the external command as a single step) from SM37 it fails and doesn't log anything to dev_cp or dev_xpg??
    Is this normally how the trace would work?
    Thanks again,
    Kye

  • WLC Internal and External DHCP

    I am currently using the Internal DHCP component within my 5508 Controller with software version 7.0.166.0.  This seems to be working fine as the Vlan Routed interface connected to it via the Dynamic Trunk Port is functioning as l have the ip-helper command setup on this specific vlan interface..
    My issue now is that we have a isolated ADSL Network which is configured off our Core 6513 but just as a Layer 2 Vlan so no traffic can be routed to other vlans.
    With our new WIFI environment which consists of the 5508 Controller and numerous 3502 AP's we wont to utilize this ADSL vlan with our new WIFI environment..  This ADSL Vlan has a dedicated Linksys Router which is currently running DHCP and assigning addresses to clients at the moment..
    What l want to do is configure the 5508 controller to use this ADSL vlan aswell but to also keep using the Linksys Router aswell for DHCP..
    I have setup a new dynamic interface and added the ADSL Vlan ID to the Trunk port of the 5508 and also setup its own SSID.  But for some reason l cannot get both the internal and External DHCP servers to work at the same time ?  If l enable DHCP Proxy option on the 5508 the internal DHCP server works and when l disable DHCP Proxy the ADSL Vlan DHCP works through the 5508 but not the internal DHCP Server ??
    Can l get both the internal and external DHCP servers to work in harmony or should l be focusing on using one method over the other ?

    Hey Scott l have just tried configuring another scope for the L2 Vlan but it doesn't seem to be working when l add the ip address of the management interface which is the internal DHCP Server to the dynamic interface of this adsl network l have setup l dont seem to get a ip address within this scope ?
    I am just wandering seeing it is just a L2 vlan without a routed interface would this be the problem and would need to set this up with the "ip helper-address" of the management interface ?
    Cheers SG

  • Internal and external facing applicaitons on same infrastructure

    I'm looking for suggestions on the best way to architect an apex production environment where you may have two or three apps open to the public and 10 or more for internal access only. All of the apps (regardless of public or private) are running on the same APEX instance, DB, app tier and web tier.
    We are using the APEX Listener on Weblogic for the app tier with an OHS webserver and Load Balancer in front of everything.
    The Load Balancer houses all of our certificates and has the ability to perform iRules to make more friendly urls.
    Our approach is to assign each app (ie https://someurl.com/apex/f?p=APPID) a static IP from the load balancer and then firewall public/private based on APPID to prevent internal only apps from being reached outside the network.
    Unfortunately the iRule friendly url rewrite isn't able to mask the APPID from the URL (https://someurl.com/apex/f?p=200) which currently allows anyone the ability to change the APPID parameter of the URL and cycle through all the apps regardless of the firewall rule in place to prevent it from being publicly accessible.
    For example, if we have the following apps deployed and the only one which is allowed open to the internet is app 100, the url rewrite isn't able to mask APPID of 100 (or the APP Alias if used).
    Publicly accessible:
    https://someurl.com/apex/f?p=100 (192.168.25.100)
    Internal only access:
    https://somedifferenturl.com/apex/f?p=200 (192.168.25.200)
    https://anotherurl.com/apex/f?p=250 (192.168.25.250)
    https://subdomain.someurl.com/apex/f?p=300 (192.168.25.300)
    I could navigate to the publicly accessible url https://someurl.com/apex/f?p=100 and change the APPID for one of (200,250,300) and still access those apps which should not be open to the internet.
    from the internet browsing directly to https://somedifferenturl.com/apex/f?p=200 or https://anotherurl.com/apex/f?p=250 or https://subdomain.someurl.com/apex/f?p=300 would all result in a page not found error since their ip's are not accessible directly from the internet.
    What is the best practice to overcome the above scenario and utilize shared infrastructure for internal and external facing applications? Is mod_rewrite my only other option to accomplish this setup and bypass the load balancer?

    Hi Jeff,
    I'm not sure if this is the ideal recommendation, but I know of a way you could block the "internal-only" applications from being accessed externally.
    1) Create a function which inspects the CGI environment variables, e.g., HTTP_HOST, HTTP_PORT, etc. Using this information, you determine if the request is emanating from an internal server name or an external server name.
    2) Create an authorization scheme which returns FALSE if the host/port/other CGI isn't what you expect.
    3) Apply this authorization scheme to every application you wish to keep from an external site.
    I know this isn't ideal, as you have to add this to every "internal-only" application. And if you forget an application, then this application suddenly becomes available on the Internet. But it's one way. If all of the applications are in the same workspace, you could define this authorization scheme in one application and subscribe to it from the other applications.
    Joel
    P.S. From SQL Commands, you can see all of the CGI environment variables at your disposal using:
    begin
    owa_util.print_cgi_env;
    end;

  • How to Setup RDS custom property when internal and external domain name space is different

    Hi All
    I am setting up RDS for customer
    My internal domain name is domain.local and my external domain is domain.com
    I came across below PowerShell cmdlets on some blogs because my internal and external name space are different
    Set-RDSessionCollectionConfiguration –CollectionName QuickSessionCollection -CustomRdpProperty “use redirection server name:i:1 `n alternate full address:s:remote.domain.com”
    In above command, remote.domain.com points to which host?
    Is it pointing to RD Session Broker
    OR
    Pointing to RD Session Host servers
    I am not sure what above command will do exactly ?
    Any help will be highly appreciated
    Thanks Best Regards Mahesh

    Hi,
    It all depends who is accessing the RDS Solution.
    If you have a large BYOD or large number of external users, it would be better to use a public certificate.
    Have a look at the following script which will simplyfy the configuration of the RDSH hosts with certificates.
    http://ryanmangansitblog.com/2014/05/20/rds-2012-rdsh-certificate-deployment-script/
    You can use a custom RDP property to hide the Session host names.
    Have a look at the following article on configuring certificates:
    http://ryanmangansitblog.com/2013/03/10/configuring-rds-2012-certificates-and-sso/
    Ryan Mangan | Ryanmangansitblog.wordpress.com | Help keep the forums tidy, if this has helped please mark it as an answer

Maybe you are looking for